INTRODUCTION TO IMPACT EVALUATION AND
RANDOMIZED CONTROL TRIALS
PRESENTATION 2: WHY RANDOMIZE?
PARTICIPATION AND REGULATORY COMPLIANCE
PROJECT LAUNCH WORKSHOP
8-9 JULY 2014. HANOI
Héctor Salazar Salame, Executive Director J-PAL SEA
Presentations Overview
1. What is evaluation? Why Evaluate?
2. Why randomize?
3. How to randomize
4. Evaluation from Start to Finish
Presentations Overview
1. What is evaluation? Why Evaluate?
2. Why randomize?
3. How to randomize
4. Evaluation from Start to Finish
I. Background
II. What is a randomized experiment?
III. Why randomize?
IV.Common criticisms and responses
Lecture Overview
I - Background
Impact: What is it?
Time
PrimaryOutcome
Impact
Intervention
How to measure impact?
Impact is defined as a comparison between:
1. the outcome some time after the program has
been introduced
2. the outcome at that same point in time had the
program not been introduced (the
”counterfactual”)
7
Impact: What is it?
Time
PrimaryOutcome
Impact
Intervention
Impact: What is it?
Time
PrimaryOutcome
Impact
Intervention
Counterfactual
• The counterfactual represents the state of
the world that program participants would
have experienced in the absence of the
program (i.e. had they not participated in
the program)
• Problem: Counterfactual cannot be
observed
• Solution: We need to “mimic” or construct
the counterfactual
Constructing the counterfactual
• Usually done by selecting a group of individuals
that did not participate in the program
• This group is usually referred to as the control
group or comparison group
• How this group is selected is a key decision in the
design of any impact evaluation
Selecting the comparison group
• Idea: Select a group that is exactly like the group of
participants in all ways except one: their exposure to
the program being evaluated
• Goal: To be able to attribute differences in outcomes
between the group of participants and the comparison
group to the program (and not to other factors)
Impact evaluation methods
1. Randomized Experiments
• Also known as:
– Random Assignment Studies
– Randomized Field Trials
– Social Experiments
– Randomized Controlled Trials (RCTs)
– Randomized Controlled Experiments
13
Impact evaluation methods
2. Non- or Quasi-Experimental Methods
a. Pre-Post
b. Simple Difference
c. Differences-in-Differences
d. Multivariate Regression
e. Statistical Matching
f. Interrupted Time Series
g. Instrumental Variables
h. Regression Discontinuity
II – What is a randomized
experiment?
The basics
Start with simple case:
• Take a sample of program applicants
• Randomly assign them to either:
 Treatment Group – is offered treatment
 Control Group - not offered treatment
(during the evaluation period)
Key advantage of experiments
Because members of the groups (treatment
and control) do not differ systematically at
the outset of the experiment,
any difference that subsequently arises
between them can be attributed to the
program rather than to other factors.
17
Evaluation of “Women as Policymakers”:
Treatment vs. Control villages at baseline
Variables
Treatment
Group
Control
Group
Difference
Female Literacy Rate 0.35 0.34
0.01
(0.01)
Number of Public Health
Facilities
0.06 0.08
-0.02
(0.02)
Tap Water 0.05 0.03
0.02
(0.02)
Number of Primary Schools 0.95 0.91
0.04
(0.08)
Number of High Schools 0.09 0.10
-0.01
(0.02)
Standard Errors in parentheses. Statistics displayed for West
Bengal
*/*/***: Statistically significant at the 10% / 5% / 1% level
Source: Chattopadhyay and Duflo (2004)
Some variations on the basics
• Assigning to multiple treatment groups
• Assigning of units other than individuals
or households
 Health Centers
 Schools
 Local Governments
 Villages
Key steps in conducting an experiment
1. Design the study carefully
2. Randomly assign people to treatment or
control
3. Collect baseline data
4. Verify that assignment looks random
5. Monitor process so that integrity of
experiment is not compromised
Key steps in conducting an experiment
(cont.)
6. Collect follow-up data for both the
treatment and control groups
7. Estimate program impacts by comparing
mean outcomes of treatment group vs.
mean outcomes of control group.
8. Assess whether program impacts are
statistically significant and practically
significant.
III – Why randomize?
Why randomize? – Conceptual Argument
If properly designed and conducted,
randomized experiments provide the most
credible method to estimate the impact of a
program
23
Why “most credible”?
Because members of the groups (treatment
and control) do not differ systematically at
the outset of the experiment,
Any difference that subsequently arises
between them can be attributed to the
program rather than to other factors.
24
Example: Balsakhi Program
Case 2: Remedial Education in IndiaCase 2: Remedial Education in India
Balsakhi Program: Background
• Implemented by Pratham, an NGO in India
• Program provided tutors (Balsakhi) to help
at-risk children with school work
• In Vadodara, the balsakhi program was run
in government primary schools in 2002-
2003
• Teachers decided which children would get
the balsakhi
Balsakhi: Outcomes
• Children were tested at the beginning of the
school year (Pretest) and at the end of the year
(Post-test)
• QUESTION: How can we estimate the impact
of the balsakhi program on test scores?
Methods to estimate impacts
• Let’s look at different ways of estimating
the impacts using the data from the schools
that got a balsakhi
1. Pre – Post (Before vs. After)
2. Simple difference
3. Difference-in-difference
4. Other non-experimental methods
5. Randomized Experiment
• Look at average
change in test scores
over the school year
for the balsakhi
children
1 - Pre-post (Before vs. After)
1 - Pre-post (Before vs. After)
• QUESTION: Under what conditions can this
difference (26.42) be interpreted as the impact
of the balsakhi program?
Average post-test score for
children with a balsakhi
51.22
Average pretest score for
children with a balsakhi
24.80
Difference 26.42
What would have happened without balsakhi?
Method 1: Before vs. After
Impact = 26.42 points?
75
50
25
0
0
2002 2003
26.42 points?
2 - Simple difference
Children who got
balsakhi
Compare test scores of…
Children who did not get
balsakhi
With
test
scores
of…
2 - Simple difference
• QUESTION: Under what conditions can this
difference (-5.05) be interpreted as the impact
of the balsakhi program?
Average score for children
with a balsakhi
51.22
Average score for children
without a balsakhi
56.27
Difference -5.05
What would have happened without balsakhi?
Method 2: Simple Comparison
Impact = -5.05 points?
75
50
25
0
0
2002 2003
-5.05 points?
3 – Difference-in-Differences
Children who got
balsakhi
Compare gains in test scores of…
Children who did not get
balsakhi
With
gains
in test
scores
of…
3 - Difference-in-differences
Pretest Post-test Difference
Average score for children
with a balsakhi
24.80 51.22 26.42
3 - Difference-in-differences
Pretest Post-test Difference
Average score for children
with a balsakhi
24.80 51.22 26.42
Average score for children
without a balsakhi
36.67 56.27 19.60
• QUESTION: Under what conditions can 6.82 be
interpreted as the impact of the balsakhi program?
3 - Difference-in-differences
Pretest Post-test Difference
Average score for children
with a balsakhi
24.80 51.22 26.42
Average score for children
without a balsakhi
36.67 56.27 19.60
Difference 6.82
• There are more sophisticated non-experimental
methods to estimate program impacts:
– Regression
– Matching
– Instrumental Variables
– Regression Discontinuity
• These methods rely on being able to “mimic” the
counterfactual under certain assumptions
• Problem: Assumptions are not testable
4 – Other Methods
• Suppose we evaluated the balsakhi program
using a randomized experiment
• QUESTION #1: What would this entail?
How would we do it?
• QUESTION #2: What would be the
advantage of using this method to evaluate
the impact of the balsakhi program?
5 – Randomized Experiment
40Source: www.theoryofchange.org
Impact of Balsakhi - Summary
Method Impact Estimate
(1) Pre-post 26.42*
(2) Simple Difference -5.05*
(3) Difference-in-Difference 6.82*
(4) Regression 1.92
*: Statistically significant at the 5% level
Impact of Balsakhi - Summary
Method Impact Estimate
(1) Pre-post 26.42*
(2) Simple Difference -5.05*
(3) Difference-in-Difference 6.82*
(4) Regression 1.92
(5) Randomized Experiment 5.87*
*: Statistically significant at the 5% level
Impact of Balsakhi - Summary
Method Impact Estimate
(1) Pre-post 26.42*
(2) Simple Difference -5.05*
(3) Difference-in-Difference 6.82*
(4) Regression 1.92
(5)Randomized Experiment 5.87*
Bottom Line: Which method we use matters!
*: Statistically significant at the 5% level
Example #2 – South Africa microfinance
Method Impact Estimate
(1) Pre-post 2384*
(2) Simple Difference 1838*
(3) Difference-in-Difference 1068*
(4) Regression 1412
(5)Randomized Experiment
*: Statistically significant at the 5% level
Example #2 – South Africa microfinance
Method Impact Estimate
(1) Pre-post 2384*
(2) Simple Difference 1838*
(3) Difference-in-Difference 1068*
(4) Regression 1412
(5)Randomized Experiment 292*
*: Statistically significant at the 5% level
Example #3 - Pratham’s Read India program
Example #3 - Pratham’s Read India program
Method Impact
(1) Pre-Post 0.60*
(2) Simple Difference -0.90*
(3) Difference-in-Differences 0.31*
(4) Regression 0.06
(5) Randomized Experiment
*: Statistically significant at the 5% level
Example #3 - Pratham’s Read India program
Method Impact
(1) Pre-Post 0.60*
(2) Simple Difference -0.90*
(3) Difference-in-Differences 0.31*
(4) Regression 0.06
(5) Randomized Experiment 0.88*
*: Statistically significant at the 5% level
IV – Conclusions
• There are many ways to estimate a
program’s impact
• This course argues in favor of one:
randomized experiments
– Conceptual argument: If properly designed and
conducted, randomized experiments provide the
most credible method to estimate the impact of
a program
– Empirical argument: Different methods can
generate different impact estimates
Conclusions - Why Randomize?
50Source: www.theoryofchange.org
• When is a RCT not possible?
• If it is possible, when would a RCT be
unnecessary?
• What are common critiques you have heard
of RCTs?
Conclusions – A few parting questions
51Source: www.theoryofchange.org
HOW TO RANDOMIZE
Up Next…

More Related Content

PPTX
3. How to Randomize
PPTX
Vicky Pelka's Training Session On Impact Evaluation
PPTX
Innovative Study Designs for Implementation Research
PPTX
Impact evaluation
PPTX
Root cause analysis
PDF
Faster Improvement with Adaptive Implementation Research
PPTX
Total Survey Error across a program of three national surveys: using a risk m...
3. How to Randomize
Vicky Pelka's Training Session On Impact Evaluation
Innovative Study Designs for Implementation Research
Impact evaluation
Root cause analysis
Faster Improvement with Adaptive Implementation Research
Total Survey Error across a program of three national surveys: using a risk m...

What's hot (20)

PPTX
Delphi in community assessment na
PDF
Critical Appraisal Skills Programme (CASP) Cohort Study Checklist Worksheet (...
PPTX
TIARA Module 4 Geoff Barnes 10302019
PPTX
Ipac 2014
PDF
Contractor Safety Beyond Compliance
PDF
Handout #1 - QIAMay4
PPT
Total Survey Error & Institutional Research: A case study of the University E...
PDF
l’outil CASP pour les études qualitatives – webinaire du Club de lecture en l...
PPTX
Sampling, Statistics and Sample Size
PDF
The art and science of managing quality
PPTX
Cedp 402 assessment lectures
PDF
Van der vleuten_-_twelve_tips_for_programmatic_assessment
PPTX
Towards Dementia-friendly Smart Home
PDF
IDEA Best Practice Chairs & Admin
PPTX
Research methodology methods and techniques 2004
PDF
Beat the odds evaluation model table
PPT
The use of models and decision support tools: UK perspective (N. Gibbens)
PPT
Designing Consumer Market Research by Fida Karim
Delphi in community assessment na
Critical Appraisal Skills Programme (CASP) Cohort Study Checklist Worksheet (...
TIARA Module 4 Geoff Barnes 10302019
Ipac 2014
Contractor Safety Beyond Compliance
Handout #1 - QIAMay4
Total Survey Error & Institutional Research: A case study of the University E...
l’outil CASP pour les études qualitatives – webinaire du Club de lecture en l...
Sampling, Statistics and Sample Size
The art and science of managing quality
Cedp 402 assessment lectures
Van der vleuten_-_twelve_tips_for_programmatic_assessment
Towards Dementia-friendly Smart Home
IDEA Best Practice Chairs & Admin
Research methodology methods and techniques 2004
Beat the odds evaluation model table
The use of models and decision support tools: UK perspective (N. Gibbens)
Designing Consumer Market Research by Fida Karim
Ad

Similar to 2. Why Randomize (20)

PPTX
Evaluation Methods
PDF
Key Issues in Impact Evaluation: A MEET and GEMNet-Health Virtual Event
PDF
IFPRI - The Problem of Impact Evaluation
PDF
Gilligan quantitative impact eval methods
PPTX
Impact evaluation an overview
PPT
Program evaluation part 2
PPTX
Evaluation Design presentation public health
PPTX
Experimental Evaluation Methods
PDF
ICAR-IFPRI : Problems of Impact Evaluation Confounding Factors and Selection ...
PPTX
1.2 Methods of EIA.pptxgrhtghrgffrgrhdfcd
PPTX
What is Evaluation
PPTX
Cadth 2015 b3 rapid reviews oral presentation cadth_11_apr2015
PDF
Ahmed ifpri impact evaluation methods_15 nov 2011
PPT
Program Evaluation 1
PDF
Getting started with a systematic review: developing your review question
PPTX
Choosing Appropriate Evaluation Methods tool
PDF
Health education and promotion programs
PPTX
COMMUNITY EVALUATION 2023.pptx
PDF
Session 1 general overview
PPT
Rogers Revised Presentation For Cali Workshop
Evaluation Methods
Key Issues in Impact Evaluation: A MEET and GEMNet-Health Virtual Event
IFPRI - The Problem of Impact Evaluation
Gilligan quantitative impact eval methods
Impact evaluation an overview
Program evaluation part 2
Evaluation Design presentation public health
Experimental Evaluation Methods
ICAR-IFPRI : Problems of Impact Evaluation Confounding Factors and Selection ...
1.2 Methods of EIA.pptxgrhtghrgffrgrhdfcd
What is Evaluation
Cadth 2015 b3 rapid reviews oral presentation cadth_11_apr2015
Ahmed ifpri impact evaluation methods_15 nov 2011
Program Evaluation 1
Getting started with a systematic review: developing your review question
Choosing Appropriate Evaluation Methods tool
Health education and promotion programs
COMMUNITY EVALUATION 2023.pptx
Session 1 general overview
Rogers Revised Presentation For Cali Workshop
Ad

More from vinhthedang (12)

PDF
Bài giảng và bài tập chọn mẫu và tính toán cỡ mẫu
PDF
8. Novels, Society and History
PDF
7. Print Culture and the Modern World
PDF
6. Work, Life and Leisure
PDF
5. The Age of Industrialisation
PDF
4. The Making of a Global World
PDF
3. Nationalism in India
PDF
2. The Nationalist Movement in Indo-China
PDF
1. The Rise of Nationalism in Europe
PPTX
4. Start to Finish Raskin
PPT
1. About J-Pal and What is Evaluation
PPTX
Survey Procedures at Mekong Development Research Institute
Bài giảng và bài tập chọn mẫu và tính toán cỡ mẫu
8. Novels, Society and History
7. Print Culture and the Modern World
6. Work, Life and Leisure
5. The Age of Industrialisation
4. The Making of a Global World
3. Nationalism in India
2. The Nationalist Movement in Indo-China
1. The Rise of Nationalism in Europe
4. Start to Finish Raskin
1. About J-Pal and What is Evaluation
Survey Procedures at Mekong Development Research Institute

Recently uploaded (20)

PPTX
Presentation on CGIAR’s Policy Innovation Program _18.08.2025 FE.pptx
PPTX
Chapter 1: Philippines constitution laws
PDF
Building Bridges (of Hope) over Our Troubled Waters_PART 1
PDF
UNEP/ UNEA Plastic Treaty Negotiations Report of Inc 5.2 Geneva
PPTX
True Fruits_ reportcccccccccccccccc.pptx
PPTX
ANALYSIS OF THE PROCLAMATION OF THE PHILIPPHINE INDEPENDENCE.pptx
PDF
An Easy Approach to Kerala Service Rules
PDF
eVerify Overview and Detailed Instructions to Set up an account
PPTX
Key Points of 2025 ORAOHRA of the CSC from CSI
PDF
PPT Item # 10 -- Proposed 2025 Tax Rate
PDF
RBI-FORM-A-By Household_Revised 2024.pdf
PPTX
CHS rollout Presentation by Abraham Lebeza.pptx
PDF
Global Intergenerational Week Impact Report
PPTX
Neurons.pptx and the family in London are you chatgpt
PPTX
Robotics_Presentation.pptxdhdrhdrrhdrhdrhdrrh
PPTX
International Tracking Project Unloading Guidance Manual V1 (1) 1.pptx
PPTX
Developing_An_Advocacy_Agenda_by_Kevin_Karuga.pptx
PPTX
Workshop introduction and objectives. SK.pptx
PPTX
Parliamentary procedure in meeting that can be use
PPT
Republic Act 9729 Climate Change Adaptation
Presentation on CGIAR’s Policy Innovation Program _18.08.2025 FE.pptx
Chapter 1: Philippines constitution laws
Building Bridges (of Hope) over Our Troubled Waters_PART 1
UNEP/ UNEA Plastic Treaty Negotiations Report of Inc 5.2 Geneva
True Fruits_ reportcccccccccccccccc.pptx
ANALYSIS OF THE PROCLAMATION OF THE PHILIPPHINE INDEPENDENCE.pptx
An Easy Approach to Kerala Service Rules
eVerify Overview and Detailed Instructions to Set up an account
Key Points of 2025 ORAOHRA of the CSC from CSI
PPT Item # 10 -- Proposed 2025 Tax Rate
RBI-FORM-A-By Household_Revised 2024.pdf
CHS rollout Presentation by Abraham Lebeza.pptx
Global Intergenerational Week Impact Report
Neurons.pptx and the family in London are you chatgpt
Robotics_Presentation.pptxdhdrhdrrhdrhdrhdrrh
International Tracking Project Unloading Guidance Manual V1 (1) 1.pptx
Developing_An_Advocacy_Agenda_by_Kevin_Karuga.pptx
Workshop introduction and objectives. SK.pptx
Parliamentary procedure in meeting that can be use
Republic Act 9729 Climate Change Adaptation

2. Why Randomize

  • 1. INTRODUCTION TO IMPACT EVALUATION AND RANDOMIZED CONTROL TRIALS PRESENTATION 2: WHY RANDOMIZE? PARTICIPATION AND REGULATORY COMPLIANCE PROJECT LAUNCH WORKSHOP 8-9 JULY 2014. HANOI Héctor Salazar Salame, Executive Director J-PAL SEA
  • 2. Presentations Overview 1. What is evaluation? Why Evaluate? 2. Why randomize? 3. How to randomize 4. Evaluation from Start to Finish
  • 3. Presentations Overview 1. What is evaluation? Why Evaluate? 2. Why randomize? 3. How to randomize 4. Evaluation from Start to Finish
  • 4. I. Background II. What is a randomized experiment? III. Why randomize? IV.Common criticisms and responses Lecture Overview
  • 6. Impact: What is it? Time PrimaryOutcome Impact Intervention
  • 7. How to measure impact? Impact is defined as a comparison between: 1. the outcome some time after the program has been introduced 2. the outcome at that same point in time had the program not been introduced (the ”counterfactual”) 7
  • 8. Impact: What is it? Time PrimaryOutcome Impact Intervention
  • 9. Impact: What is it? Time PrimaryOutcome Impact Intervention
  • 10. Counterfactual • The counterfactual represents the state of the world that program participants would have experienced in the absence of the program (i.e. had they not participated in the program) • Problem: Counterfactual cannot be observed • Solution: We need to “mimic” or construct the counterfactual
  • 11. Constructing the counterfactual • Usually done by selecting a group of individuals that did not participate in the program • This group is usually referred to as the control group or comparison group • How this group is selected is a key decision in the design of any impact evaluation
  • 12. Selecting the comparison group • Idea: Select a group that is exactly like the group of participants in all ways except one: their exposure to the program being evaluated • Goal: To be able to attribute differences in outcomes between the group of participants and the comparison group to the program (and not to other factors)
  • 13. Impact evaluation methods 1. Randomized Experiments • Also known as: – Random Assignment Studies – Randomized Field Trials – Social Experiments – Randomized Controlled Trials (RCTs) – Randomized Controlled Experiments 13
  • 14. Impact evaluation methods 2. Non- or Quasi-Experimental Methods a. Pre-Post b. Simple Difference c. Differences-in-Differences d. Multivariate Regression e. Statistical Matching f. Interrupted Time Series g. Instrumental Variables h. Regression Discontinuity
  • 15. II – What is a randomized experiment?
  • 16. The basics Start with simple case: • Take a sample of program applicants • Randomly assign them to either:  Treatment Group – is offered treatment  Control Group - not offered treatment (during the evaluation period)
  • 17. Key advantage of experiments Because members of the groups (treatment and control) do not differ systematically at the outset of the experiment, any difference that subsequently arises between them can be attributed to the program rather than to other factors. 17
  • 18. Evaluation of “Women as Policymakers”: Treatment vs. Control villages at baseline Variables Treatment Group Control Group Difference Female Literacy Rate 0.35 0.34 0.01 (0.01) Number of Public Health Facilities 0.06 0.08 -0.02 (0.02) Tap Water 0.05 0.03 0.02 (0.02) Number of Primary Schools 0.95 0.91 0.04 (0.08) Number of High Schools 0.09 0.10 -0.01 (0.02) Standard Errors in parentheses. Statistics displayed for West Bengal */*/***: Statistically significant at the 10% / 5% / 1% level Source: Chattopadhyay and Duflo (2004)
  • 19. Some variations on the basics • Assigning to multiple treatment groups • Assigning of units other than individuals or households  Health Centers  Schools  Local Governments  Villages
  • 20. Key steps in conducting an experiment 1. Design the study carefully 2. Randomly assign people to treatment or control 3. Collect baseline data 4. Verify that assignment looks random 5. Monitor process so that integrity of experiment is not compromised
  • 21. Key steps in conducting an experiment (cont.) 6. Collect follow-up data for both the treatment and control groups 7. Estimate program impacts by comparing mean outcomes of treatment group vs. mean outcomes of control group. 8. Assess whether program impacts are statistically significant and practically significant.
  • 22. III – Why randomize?
  • 23. Why randomize? – Conceptual Argument If properly designed and conducted, randomized experiments provide the most credible method to estimate the impact of a program 23
  • 24. Why “most credible”? Because members of the groups (treatment and control) do not differ systematically at the outset of the experiment, Any difference that subsequently arises between them can be attributed to the program rather than to other factors. 24
  • 25. Example: Balsakhi Program Case 2: Remedial Education in IndiaCase 2: Remedial Education in India
  • 26. Balsakhi Program: Background • Implemented by Pratham, an NGO in India • Program provided tutors (Balsakhi) to help at-risk children with school work • In Vadodara, the balsakhi program was run in government primary schools in 2002- 2003 • Teachers decided which children would get the balsakhi
  • 27. Balsakhi: Outcomes • Children were tested at the beginning of the school year (Pretest) and at the end of the year (Post-test) • QUESTION: How can we estimate the impact of the balsakhi program on test scores?
  • 28. Methods to estimate impacts • Let’s look at different ways of estimating the impacts using the data from the schools that got a balsakhi 1. Pre – Post (Before vs. After) 2. Simple difference 3. Difference-in-difference 4. Other non-experimental methods 5. Randomized Experiment
  • 29. • Look at average change in test scores over the school year for the balsakhi children 1 - Pre-post (Before vs. After)
  • 30. 1 - Pre-post (Before vs. After) • QUESTION: Under what conditions can this difference (26.42) be interpreted as the impact of the balsakhi program? Average post-test score for children with a balsakhi 51.22 Average pretest score for children with a balsakhi 24.80 Difference 26.42
  • 31. What would have happened without balsakhi? Method 1: Before vs. After Impact = 26.42 points? 75 50 25 0 0 2002 2003 26.42 points?
  • 32. 2 - Simple difference Children who got balsakhi Compare test scores of… Children who did not get balsakhi With test scores of…
  • 33. 2 - Simple difference • QUESTION: Under what conditions can this difference (-5.05) be interpreted as the impact of the balsakhi program? Average score for children with a balsakhi 51.22 Average score for children without a balsakhi 56.27 Difference -5.05
  • 34. What would have happened without balsakhi? Method 2: Simple Comparison Impact = -5.05 points? 75 50 25 0 0 2002 2003 -5.05 points?
  • 35. 3 – Difference-in-Differences Children who got balsakhi Compare gains in test scores of… Children who did not get balsakhi With gains in test scores of…
  • 36. 3 - Difference-in-differences Pretest Post-test Difference Average score for children with a balsakhi 24.80 51.22 26.42
  • 37. 3 - Difference-in-differences Pretest Post-test Difference Average score for children with a balsakhi 24.80 51.22 26.42 Average score for children without a balsakhi 36.67 56.27 19.60
  • 38. • QUESTION: Under what conditions can 6.82 be interpreted as the impact of the balsakhi program? 3 - Difference-in-differences Pretest Post-test Difference Average score for children with a balsakhi 24.80 51.22 26.42 Average score for children without a balsakhi 36.67 56.27 19.60 Difference 6.82
  • 39. • There are more sophisticated non-experimental methods to estimate program impacts: – Regression – Matching – Instrumental Variables – Regression Discontinuity • These methods rely on being able to “mimic” the counterfactual under certain assumptions • Problem: Assumptions are not testable 4 – Other Methods
  • 40. • Suppose we evaluated the balsakhi program using a randomized experiment • QUESTION #1: What would this entail? How would we do it? • QUESTION #2: What would be the advantage of using this method to evaluate the impact of the balsakhi program? 5 – Randomized Experiment 40Source: www.theoryofchange.org
  • 41. Impact of Balsakhi - Summary Method Impact Estimate (1) Pre-post 26.42* (2) Simple Difference -5.05* (3) Difference-in-Difference 6.82* (4) Regression 1.92 *: Statistically significant at the 5% level
  • 42. Impact of Balsakhi - Summary Method Impact Estimate (1) Pre-post 26.42* (2) Simple Difference -5.05* (3) Difference-in-Difference 6.82* (4) Regression 1.92 (5) Randomized Experiment 5.87* *: Statistically significant at the 5% level
  • 43. Impact of Balsakhi - Summary Method Impact Estimate (1) Pre-post 26.42* (2) Simple Difference -5.05* (3) Difference-in-Difference 6.82* (4) Regression 1.92 (5)Randomized Experiment 5.87* Bottom Line: Which method we use matters! *: Statistically significant at the 5% level
  • 44. Example #2 – South Africa microfinance Method Impact Estimate (1) Pre-post 2384* (2) Simple Difference 1838* (3) Difference-in-Difference 1068* (4) Regression 1412 (5)Randomized Experiment *: Statistically significant at the 5% level
  • 45. Example #2 – South Africa microfinance Method Impact Estimate (1) Pre-post 2384* (2) Simple Difference 1838* (3) Difference-in-Difference 1068* (4) Regression 1412 (5)Randomized Experiment 292* *: Statistically significant at the 5% level
  • 46. Example #3 - Pratham’s Read India program
  • 47. Example #3 - Pratham’s Read India program Method Impact (1) Pre-Post 0.60* (2) Simple Difference -0.90* (3) Difference-in-Differences 0.31* (4) Regression 0.06 (5) Randomized Experiment *: Statistically significant at the 5% level
  • 48. Example #3 - Pratham’s Read India program Method Impact (1) Pre-Post 0.60* (2) Simple Difference -0.90* (3) Difference-in-Differences 0.31* (4) Regression 0.06 (5) Randomized Experiment 0.88* *: Statistically significant at the 5% level
  • 50. • There are many ways to estimate a program’s impact • This course argues in favor of one: randomized experiments – Conceptual argument: If properly designed and conducted, randomized experiments provide the most credible method to estimate the impact of a program – Empirical argument: Different methods can generate different impact estimates Conclusions - Why Randomize? 50Source: www.theoryofchange.org
  • 51. • When is a RCT not possible? • If it is possible, when would a RCT be unnecessary? • What are common critiques you have heard of RCTs? Conclusions – A few parting questions 51Source: www.theoryofchange.org