SlideShare a Scribd company logo
Practitioners’ Expectations on
Automated Fault Localization
Pavneet Singh Kochhar*, Xin Xia+, David Lo*, Shanping Li+
*Singapore Management University
+Zhejiang University
The International Symposium on Software Testing and Analysis (ISSTA)
Too many bugs!
• Many projects receive large numbers of bug reports.
• Large number of bug reports can overwhelm developers.
- Mozilla developer - “Everyday, almost 300 bugs appear that need
triaging. This is far too much for only the Mozilla programmers to
handle” *
What have researchers proposed to overcome this issue?
*J. Anvik, L. Hiew, and G. C. Murphy, “Coping with an open bug
repository,” in ETX, pp. 35–39, 2005
2/31
Fault Localization
Thousands of Source Code Files
Find the buggy files/
methods/statements/
blocks
3/31
------>
GOAL:
How Fault Localization Works
4/31
Bug Reports Test Cases
Fault Localization Techniques
Information Retrieval-Based, Slicing, Spectrum-Based etc.
Statements Methods Classes
Fault Localization
What are the expectations of practitioners
on fault localization?
What factors affect adoption of fault
localization tools?
What are the thresholds for adoption?
5/31
Our Study
Practitioners Expectations
6/31
Survey Literature
Review
Practitioner
Survey
7/31
Practitioners Survey
• Multi-pronged strategy:
• Our contacts in IT industry
• Email 3300 practitioners on
• We receive 403 responses.
8/31
Survey Demographics
• 386 responses
• 33 countries
• Job profile
• Software Dev – 80.83%
• Software Testing – 30.05%
• Project Management – 17.10%
• Professional – 78.13%, Open-source – 44.24%
9/31
RQ1: Importance of Fault Localization
10/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
All Dev Test PM ExpLow ExpMed ExpHigh OS Prof
Ratings
Demographics
Essential Worthwhile Unimportant Unwise
RQ1: Importance of Fault Localization
11/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
All Dev Test PM ExpLow ExpMed ExpHigh OS Prof
Ratings
Demographics
Essential Worthwhile Unimportant Unwise
Fisher’s Exact Test = p-values < 0.05
RQ1: Importance of Fault Localization
Why “Unimportant” or “Unwise”
• Can’t deal with difficult bugs
- “I’m well aware of what static analysis can do and very few
hard bugs would be solved with it.”
• No rationale
- “I doubt any automated software can explain the reason for
things such as broken backwards compatibility, unclear
documentation etc.”
• Status quo
- “I don’t think personally I would pay for it, because for my cases
usual stack trace is over than enough”
12/31
RQ2: Availability of Debugging Data
13/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc
Ratings
Debugging Hints Available
All the time Sometimes Rarely Never
RQ2: Availability of Debugging Data
14/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc
Ratings
Debugging Hints Available
All the time Sometimes Rarely Never
>70% respondents mention availability of test cases
RQ2: Availability of Debugging Data
15/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc
Ratings
Debugging Hints Available
All the time Sometimes Rarely Never
>80% respondents mention availability of bug reports
RQ3: Preferred Granularity Level
16/31
20.21%
26.42%
51.81%
44.30%
50.00%
0%
20%
40%
60%
Component Class Method Block Statement
PercentageofRespondents
Preferred Granularity Level
RQ4: Minimum Success Criterion
17/31
Position of the buggy element in returned list
9.43%
73.58%
15.09%
1.35% 0.54%
0%
25%
50%
75%
100%
Top 1 Top 5 Top 10 Top 20 Top 50
PercentageofRespondents
Minimum Success Criterion
RQ5: Trustworthiness
18/31
Proportion of times a technique works.
0%
25%
50%
75%
100%
5% 20% 50% 75% 90% 100%
SatisfactionRate
Minimum Success Rate
RQ6: Scalability
19/31
Program sizes a technique can work on.
0%
25%
50%
75%
100%
1-100 1-1000 1-10,000 1-100,000 1-1000,000
SatisfactionRate
Minimum Program Size
RQ7: Efficiency
20/31
Time taken to produce the results.
0%
25%
50%
75%
100%
< 1 seconds < 1 minute < 30 minutes < 1 hour < 1 day
SatisfactionRate
Maximum Runtime
RQ8: Willingness to Adopt
21/31
• > 98% willing to adopt a trustworthy, scalable and efficient
fault localization technique.
• Unwilling
- Resistance to Change
“Since I already have one and to use another would require
training time and time to get used to it”
- More information needed
“Would it be open source? Would it work with my main
programming language? Would it work with distributed
environments?”
- Disbelief of possibility of success
“I don’t think you can do it.”
RQ9: Other Factors (Hypotheses)
22/31
• Rationale
- An automated debugging tool must provide a rationale
why some program locations are marked as suspicious.
- I will *still adopt* an efficient, scalable, and trustworthy
automated debugging tool, even if it cannot provide
rationales.
• IDE Integration
- An automated debugging tool must be integrated well to
my favourite IDE.
- I will *still adopt* a an efficient, scalable, and trustworthy
automated debugging tool, even if it is not integrated well
to my favourite IDE.
RQ9: Other Factors
23/31
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
Rationale Adoption w/o
Rationale
IDE Adoption w/o
IDE
Ratings
Statements
Strongly Agree Agree Neutral Disagree Strongly Disagree
RQ9: Other Factors
24/31
• Rationale
- False Positives
“False positives are worst than false negatives in my opinion”
- Rationale for buggy code
“Because to make a decisions about bug fixing I want to
*exactly* know why the automated tool “thinks” that the code
have a bug.”
• IDE Integration
- Extra steps needed
“No integration means extra steps which means testing will be
more cumbersome and hence less used.”
- Strong Reliance on IDE
“IDE is our environment. If I can’t add something into my
environment, it’s useless.”
Literature
Review
25/31
Literature Review
26/31
• Papers published in last 5 years (2011-2015)
- ICSE (417) ---> 2
- FSE/ESEC-FSE (255) ---> 5
- ISSTA (169) ---> 3
- TSE (350) ---> 2
- TOSEM (137) ---> 4
• Included papers
- Spectrum-Based fault localization, Information-retrieval-
Based etc.
• Excluded papers
- Automatic repair, empirical study on debugging, bug
prediction, bug detection etc.
16
papers
Literature Review
Factor Type Papers
Debugging Data
Specification -
Test Cases [4], [5], [24], [29],
[35], [40], [44], [55],
[57], [59]
Bug Reports [16], [19], [24], [52],
[56], [60]
Granularity
Method [24], [52]
Statement [4], [5], [29], [35],
[44], [55], [57], [59]
Basic Block [16]
Other [19], [40], [56], [60]
27/31
Literature Review
Factor Satisfaction Rate Papers
Success Rate
90% (90%) -
75% (75%) -
50% (50%) [16], [19], [35], [40], [52], [56],
[59], [60]
? [4], [5], [29], [55], [57]
Scalability
90% (≥1M LOC) [29], [52]
75% (≥100,000 LOC) [16], [24], [56], [59], [60]
50% (≥10,000 LOC) [4], [5], [35], [40], [44], [55], [57]
? [19]
Efficiency
90% (<1 minute) [4], [24], [40], [44], [56]
? [16], [19], [29], [52], [57], [60]
28/31
Literature Review
Factor Support? Papers
Rationale Yes [29], [44]
IDE
Integration
Yes -
29/31
Key Takeaways
Large demand for fault localization
 >97% mention “Essential” or “Worthwhile”
High adoption barrier
 Satisfy 75% of practitioners; successful results in Top 5,
works 75% of time; ≥100,000 LOC; takes <1 minute.
Current techniques can’t satisfy 75% of respondents.
Techniques that satisfy 50% of respondents work on
coarse granularity (class or file).
Rationale and IDE Integration are important.
30/31
Future Work
Develop fault localization techniques to bring
current state-of-research closer to practitioners
expectations.
Systematic Literature Review (SLR)
31/31
Thank You!
Pavneet Singh Kochhar
kochharps.wix.com/pavneet
Email: kochharps.2012@smu.edu.sg
Conclusion
386 practitioners surveyed from 33 countries.
Test cases and bug reports are often available.
Preferred granularity - Method & Statement
Preferred Success Criterion – Top 5.
Different satisfaction rates for trustworthiness,
scalability and efficiency.
Rationale and IDE Integration are important.
33/30

More Related Content

PPTX
Santa Barbara Agile: Exploratory Testing Explained and Experienced
PPTX
Exploratory testing workshop
PPTX
Software Analytics: The Dark Side and the Test Side
PPTX
First steps in testing analytics: Does test code quality matter?
PDF
A Taste of Exploratory Testing
PPTX
Software Analytics
PPTX
Fact or Fiction? What Software Analytics Can Do For Us
PPTX
Writing acceptable patches: an empirical study of open source project patches
Santa Barbara Agile: Exploratory Testing Explained and Experienced
Exploratory testing workshop
Software Analytics: The Dark Side and the Test Side
First steps in testing analytics: Does test code quality matter?
A Taste of Exploratory Testing
Software Analytics
Fact or Fiction? What Software Analytics Can Do For Us
Writing acceptable patches: an empirical study of open source project patches

What's hot (20)

PDF
Markus Gartner - Alternative Paths for Self-Education in Software Testing - E...
PDF
[Thao Vo] Deadly Traps of Automation Testing
PDF
Your Tests are Lying to You - Improving your Testing by Testing What Really M...
PDF
Evaluation and User Study in HCI
PDF
FutureOfTesting2008
PPTX
Uncertainty Quantification in Complex Physical Systems. (An Inroduction)
PDF
Exploratory Testing
PPTX
How to think smarter about software development
PDF
A taste of Exploratory Testing
PPTX
Measuring and Comparing the Reliability of the Structured Walkthrough Evaluat...
PDF
[Paul Holland] Trends in Software Testing
PDF
Exploratory Testing: Make It Part of Your Test Strategy
PDF
Exploratory Testing in an Agile Context
PDF
How Did I Miss That Bug? Managing Cognitive Bias in Testing
PPTX
The Axioms of Testing
PPT
Nii shonan-meeting-gsrm-20141021 - コピー
PDF
Bottom-up Adoption of Continuous Delivery in a Stage-gate Managed Software Or...
PDF
Mindful Metrics (QAotHW 2018)
PPTX
OmniTestingConf: Taking Test Automation to the Next Level
PDF
Strategies to Avoid Test Fixture Smells durin Software Evolution
Markus Gartner - Alternative Paths for Self-Education in Software Testing - E...
[Thao Vo] Deadly Traps of Automation Testing
Your Tests are Lying to You - Improving your Testing by Testing What Really M...
Evaluation and User Study in HCI
FutureOfTesting2008
Uncertainty Quantification in Complex Physical Systems. (An Inroduction)
Exploratory Testing
How to think smarter about software development
A taste of Exploratory Testing
Measuring and Comparing the Reliability of the Structured Walkthrough Evaluat...
[Paul Holland] Trends in Software Testing
Exploratory Testing: Make It Part of Your Test Strategy
Exploratory Testing in an Agile Context
How Did I Miss That Bug? Managing Cognitive Bias in Testing
The Axioms of Testing
Nii shonan-meeting-gsrm-20141021 - コピー
Bottom-up Adoption of Continuous Delivery in a Stage-gate Managed Software Or...
Mindful Metrics (QAotHW 2018)
OmniTestingConf: Taking Test Automation to the Next Level
Strategies to Avoid Test Fixture Smells durin Software Evolution
Ad

Similar to Practitioners’ Expectations on Automated Fault Localization (20)

PDF
How to Actually DO High-volume Automated Testing
PDF
IEEE 1633 Recommended Practices for Reliable Software
PDF
Can we induce change with what we measure?
PDF
TLC2018 Thomas Haver: The Automation Firehose - Be Strategic and Tactical
DOCX
Annotated Bibliography .Guidelines Annotated Bibliograph.docx
PPT
Testing 2 - Thinking Like A Tester
PDF
Experience Sharing on School Pentest Project
PDF
Software testing
PDF
Enabling Automated Software Testing with Artificial Intelligence
PDF
The Automation Firehose: Be Strategic & Tactical With Your Mobile & Web Testing
PPTX
A New Model for Testing
PDF
Software testing
PDF
Testing: an Introduction and Panorama
PPTX
How to Guarantee Continuous Value from your Test Automation
PPTX
Statistics in the age of data science, issues you can not ignore
PPTX
DockerCon SF 2019 - TDD is Dead
PPTX
New Model Testing: A New Test Process and Tool
ODP
Search Solutions 2015: Towards a new model of search relevance testing
PDF
Software Defects and SW Reliability Assessment
PPT
Design testabilty
How to Actually DO High-volume Automated Testing
IEEE 1633 Recommended Practices for Reliable Software
Can we induce change with what we measure?
TLC2018 Thomas Haver: The Automation Firehose - Be Strategic and Tactical
Annotated Bibliography .Guidelines Annotated Bibliograph.docx
Testing 2 - Thinking Like A Tester
Experience Sharing on School Pentest Project
Software testing
Enabling Automated Software Testing with Artificial Intelligence
The Automation Firehose: Be Strategic & Tactical With Your Mobile & Web Testing
A New Model for Testing
Software testing
Testing: an Introduction and Panorama
How to Guarantee Continuous Value from your Test Automation
Statistics in the age of data science, issues you can not ignore
DockerCon SF 2019 - TDD is Dead
New Model Testing: A New Test Process and Tool
Search Solutions 2015: Towards a new model of search relevance testing
Software Defects and SW Reliability Assessment
Design testabilty
Ad

More from Pavneet Singh Kochhar (13)

PPTX
Mining Testing Questions on Stack Overflow
PPTX
Cataloging GitHub Repositories
PPTX
An Exploratory Study of Functionality and Learning Resources of WebAPIs on Pr...
PPTX
Revisiting Assert Use in GitHub Projects
PPTX
A Large Scale Study of Multiple Programming Languages and Code Quality
PPTX
Adoption of Software Testing in Open Source Projects - A Preliminary Study on...
PPTX
Understanding the Test Automation Culture of App Developers
PPTX
Code Coverage and Test Suite Effectiveness: Empirical Study with Real Bugs in...
PPTX
An Empirical Study on the Adequacy of Testing in Open Source Projects
PPTX
Potential Biases in Bug Localization: Do They Matter?
PPTX
It’s Not a Bug, It’s a Feature: Does Misclassification Affect Bug Localization?
PPTX
Automatic Fine-Grained Issue Report Reclassification
PPTX
An Empirical Study of Adoption of Software Testing in Open Source Projects
Mining Testing Questions on Stack Overflow
Cataloging GitHub Repositories
An Exploratory Study of Functionality and Learning Resources of WebAPIs on Pr...
Revisiting Assert Use in GitHub Projects
A Large Scale Study of Multiple Programming Languages and Code Quality
Adoption of Software Testing in Open Source Projects - A Preliminary Study on...
Understanding the Test Automation Culture of App Developers
Code Coverage and Test Suite Effectiveness: Empirical Study with Real Bugs in...
An Empirical Study on the Adequacy of Testing in Open Source Projects
Potential Biases in Bug Localization: Do They Matter?
It’s Not a Bug, It’s a Feature: Does Misclassification Affect Bug Localization?
Automatic Fine-Grained Issue Report Reclassification
An Empirical Study of Adoption of Software Testing in Open Source Projects

Recently uploaded (20)

PPTX
ai tools demonstartion for schools and inter college
PPTX
Agentic AI : A Practical Guide. Undersating, Implementing and Scaling Autono...
PPTX
L1 - Introduction to python Backend.pptx
PPTX
CHAPTER 2 - PM Management and IT Context
PDF
Digital Strategies for Manufacturing Companies
PDF
System and Network Administraation Chapter 3
PDF
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
PPTX
Operating system designcfffgfgggggggvggggggggg
PDF
How to Migrate SBCGlobal Email to Yahoo Easily
PPTX
history of c programming in notes for students .pptx
PDF
Adobe Premiere Pro 2025 (v24.5.0.057) Crack free
PDF
Flood Susceptibility Mapping Using Image-Based 2D-CNN Deep Learnin. Overview ...
PDF
System and Network Administration Chapter 2
PDF
Navsoft: AI-Powered Business Solutions & Custom Software Development
PDF
Upgrade and Innovation Strategies for SAP ERP Customers
PPTX
Odoo POS Development Services by CandidRoot Solutions
PDF
Why TechBuilder is the Future of Pickup and Delivery App Development (1).pdf
PDF
Odoo Companies in India – Driving Business Transformation.pdf
PDF
Audit Checklist Design Aligning with ISO, IATF, and Industry Standards — Omne...
PDF
Addressing The Cult of Project Management Tools-Why Disconnected Work is Hold...
ai tools demonstartion for schools and inter college
Agentic AI : A Practical Guide. Undersating, Implementing and Scaling Autono...
L1 - Introduction to python Backend.pptx
CHAPTER 2 - PM Management and IT Context
Digital Strategies for Manufacturing Companies
System and Network Administraation Chapter 3
Internet Downloader Manager (IDM) Crack 6.42 Build 42 Updates Latest 2025
Operating system designcfffgfgggggggvggggggggg
How to Migrate SBCGlobal Email to Yahoo Easily
history of c programming in notes for students .pptx
Adobe Premiere Pro 2025 (v24.5.0.057) Crack free
Flood Susceptibility Mapping Using Image-Based 2D-CNN Deep Learnin. Overview ...
System and Network Administration Chapter 2
Navsoft: AI-Powered Business Solutions & Custom Software Development
Upgrade and Innovation Strategies for SAP ERP Customers
Odoo POS Development Services by CandidRoot Solutions
Why TechBuilder is the Future of Pickup and Delivery App Development (1).pdf
Odoo Companies in India – Driving Business Transformation.pdf
Audit Checklist Design Aligning with ISO, IATF, and Industry Standards — Omne...
Addressing The Cult of Project Management Tools-Why Disconnected Work is Hold...

Practitioners’ Expectations on Automated Fault Localization

  • 1. Practitioners’ Expectations on Automated Fault Localization Pavneet Singh Kochhar*, Xin Xia+, David Lo*, Shanping Li+ *Singapore Management University +Zhejiang University The International Symposium on Software Testing and Analysis (ISSTA)
  • 2. Too many bugs! • Many projects receive large numbers of bug reports. • Large number of bug reports can overwhelm developers. - Mozilla developer - “Everyday, almost 300 bugs appear that need triaging. This is far too much for only the Mozilla programmers to handle” * What have researchers proposed to overcome this issue? *J. Anvik, L. Hiew, and G. C. Murphy, “Coping with an open bug repository,” in ETX, pp. 35–39, 2005 2/31
  • 3. Fault Localization Thousands of Source Code Files Find the buggy files/ methods/statements/ blocks 3/31 ------> GOAL:
  • 4. How Fault Localization Works 4/31 Bug Reports Test Cases Fault Localization Techniques Information Retrieval-Based, Slicing, Spectrum-Based etc. Statements Methods Classes
  • 5. Fault Localization What are the expectations of practitioners on fault localization? What factors affect adoption of fault localization tools? What are the thresholds for adoption? 5/31
  • 8. Practitioners Survey • Multi-pronged strategy: • Our contacts in IT industry • Email 3300 practitioners on • We receive 403 responses. 8/31
  • 9. Survey Demographics • 386 responses • 33 countries • Job profile • Software Dev – 80.83% • Software Testing – 30.05% • Project Management – 17.10% • Professional – 78.13%, Open-source – 44.24% 9/31
  • 10. RQ1: Importance of Fault Localization 10/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% All Dev Test PM ExpLow ExpMed ExpHigh OS Prof Ratings Demographics Essential Worthwhile Unimportant Unwise
  • 11. RQ1: Importance of Fault Localization 11/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% All Dev Test PM ExpLow ExpMed ExpHigh OS Prof Ratings Demographics Essential Worthwhile Unimportant Unwise Fisher’s Exact Test = p-values < 0.05
  • 12. RQ1: Importance of Fault Localization Why “Unimportant” or “Unwise” • Can’t deal with difficult bugs - “I’m well aware of what static analysis can do and very few hard bugs would be solved with it.” • No rationale - “I doubt any automated software can explain the reason for things such as broken backwards compatibility, unclear documentation etc.” • Status quo - “I don’t think personally I would pay for it, because for my cases usual stack trace is over than enough” 12/31
  • 13. RQ2: Availability of Debugging Data 13/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc Ratings Debugging Hints Available All the time Sometimes Rarely Never
  • 14. RQ2: Availability of Debugging Data 14/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc Ratings Debugging Hints Available All the time Sometimes Rarely Never >70% respondents mention availability of test cases
  • 15. RQ2: Availability of Debugging Data 15/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Math-Spec Text-Spec One-Test Multi-Tests Suc-Tests Text-Desc Ratings Debugging Hints Available All the time Sometimes Rarely Never >80% respondents mention availability of bug reports
  • 16. RQ3: Preferred Granularity Level 16/31 20.21% 26.42% 51.81% 44.30% 50.00% 0% 20% 40% 60% Component Class Method Block Statement PercentageofRespondents Preferred Granularity Level
  • 17. RQ4: Minimum Success Criterion 17/31 Position of the buggy element in returned list 9.43% 73.58% 15.09% 1.35% 0.54% 0% 25% 50% 75% 100% Top 1 Top 5 Top 10 Top 20 Top 50 PercentageofRespondents Minimum Success Criterion
  • 18. RQ5: Trustworthiness 18/31 Proportion of times a technique works. 0% 25% 50% 75% 100% 5% 20% 50% 75% 90% 100% SatisfactionRate Minimum Success Rate
  • 19. RQ6: Scalability 19/31 Program sizes a technique can work on. 0% 25% 50% 75% 100% 1-100 1-1000 1-10,000 1-100,000 1-1000,000 SatisfactionRate Minimum Program Size
  • 20. RQ7: Efficiency 20/31 Time taken to produce the results. 0% 25% 50% 75% 100% < 1 seconds < 1 minute < 30 minutes < 1 hour < 1 day SatisfactionRate Maximum Runtime
  • 21. RQ8: Willingness to Adopt 21/31 • > 98% willing to adopt a trustworthy, scalable and efficient fault localization technique. • Unwilling - Resistance to Change “Since I already have one and to use another would require training time and time to get used to it” - More information needed “Would it be open source? Would it work with my main programming language? Would it work with distributed environments?” - Disbelief of possibility of success “I don’t think you can do it.”
  • 22. RQ9: Other Factors (Hypotheses) 22/31 • Rationale - An automated debugging tool must provide a rationale why some program locations are marked as suspicious. - I will *still adopt* an efficient, scalable, and trustworthy automated debugging tool, even if it cannot provide rationales. • IDE Integration - An automated debugging tool must be integrated well to my favourite IDE. - I will *still adopt* a an efficient, scalable, and trustworthy automated debugging tool, even if it is not integrated well to my favourite IDE.
  • 23. RQ9: Other Factors 23/31 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% Rationale Adoption w/o Rationale IDE Adoption w/o IDE Ratings Statements Strongly Agree Agree Neutral Disagree Strongly Disagree
  • 24. RQ9: Other Factors 24/31 • Rationale - False Positives “False positives are worst than false negatives in my opinion” - Rationale for buggy code “Because to make a decisions about bug fixing I want to *exactly* know why the automated tool “thinks” that the code have a bug.” • IDE Integration - Extra steps needed “No integration means extra steps which means testing will be more cumbersome and hence less used.” - Strong Reliance on IDE “IDE is our environment. If I can’t add something into my environment, it’s useless.”
  • 26. Literature Review 26/31 • Papers published in last 5 years (2011-2015) - ICSE (417) ---> 2 - FSE/ESEC-FSE (255) ---> 5 - ISSTA (169) ---> 3 - TSE (350) ---> 2 - TOSEM (137) ---> 4 • Included papers - Spectrum-Based fault localization, Information-retrieval- Based etc. • Excluded papers - Automatic repair, empirical study on debugging, bug prediction, bug detection etc. 16 papers
  • 27. Literature Review Factor Type Papers Debugging Data Specification - Test Cases [4], [5], [24], [29], [35], [40], [44], [55], [57], [59] Bug Reports [16], [19], [24], [52], [56], [60] Granularity Method [24], [52] Statement [4], [5], [29], [35], [44], [55], [57], [59] Basic Block [16] Other [19], [40], [56], [60] 27/31
  • 28. Literature Review Factor Satisfaction Rate Papers Success Rate 90% (90%) - 75% (75%) - 50% (50%) [16], [19], [35], [40], [52], [56], [59], [60] ? [4], [5], [29], [55], [57] Scalability 90% (≥1M LOC) [29], [52] 75% (≥100,000 LOC) [16], [24], [56], [59], [60] 50% (≥10,000 LOC) [4], [5], [35], [40], [44], [55], [57] ? [19] Efficiency 90% (<1 minute) [4], [24], [40], [44], [56] ? [16], [19], [29], [52], [57], [60] 28/31
  • 29. Literature Review Factor Support? Papers Rationale Yes [29], [44] IDE Integration Yes - 29/31
  • 30. Key Takeaways Large demand for fault localization  >97% mention “Essential” or “Worthwhile” High adoption barrier  Satisfy 75% of practitioners; successful results in Top 5, works 75% of time; ≥100,000 LOC; takes <1 minute. Current techniques can’t satisfy 75% of respondents. Techniques that satisfy 50% of respondents work on coarse granularity (class or file). Rationale and IDE Integration are important. 30/31
  • 31. Future Work Develop fault localization techniques to bring current state-of-research closer to practitioners expectations. Systematic Literature Review (SLR) 31/31
  • 32. Thank You! Pavneet Singh Kochhar kochharps.wix.com/pavneet Email: kochharps.2012@smu.edu.sg
  • 33. Conclusion 386 practitioners surveyed from 33 countries. Test cases and bug reports are often available. Preferred granularity - Method & Statement Preferred Success Criterion – Top 5. Different satisfaction rates for trustworthiness, scalability and efficiency. Rationale and IDE Integration are important. 33/30

Editor's Notes

  • #3: 22 and above
  • #5: Based
  • #13: 22
  • #17: Going down to statement does not matter much.
  • #18: Minimum success – before they find it acceptable. Our respondents not practitioners.
  • #19: Proportion
  • #20: Overlap caption
  • #21: Explain well=> what did you ask developers, how do you plot this graph from their responses
  • #27: Major conferences and journals not top. Why not bug prediction, bug detection
  • #30: Some form of rationale Read about these rationales Unclear whether the rationales help or not
  • #34: Add future work