SlideShare a Scribd company logo
Data Collection

List of Data Points

  Gerrit Klaschke
What is covered?
   Suggested Mandatory Data
   Optional Data
   Data for Model Calibrations
   Goal-Question-Metrics Approach
Data Collection – What to Collect?
   This powerpoint lists what project data needs to be
    collected for different purposes.
   First part lists mandatory data and suggested optional
    data.
   Second part lists data needed for model calibration
   Last part: we categorize the data using the GQM
    approach.
        After setting a goal and questions, the GQM approach defines
         the metrics and data points which are needed to be collected.
         These lists are also synchronized with sources such as ISBSG
         Database, USC COCOMO Project Database and QSM data
         collection documents available on the Internet
Suggested Mandatory Data
   Project name including version/
    increment/ release if applicable
   Project manager name and contact
    information
   Company/department information
   Submitter name and contact information
   Submission date
Suggested Mandatory Data
   Estimation model used including the model parameter
    used
   Project type/domain
   Project lifecycle
   Project standard – if applicable
   Equivalent SLOC or other sizing depending on the model
   Counting convention used, e.g. logical or physical SLOC
    and counting technique used e.g. counting tool
   Estimated Effort
   Estimated Schedule
Suggested Mandatory Data
   Actual Effort (person months)
   Actual Schedule (months)
   Project memo field for free text
   Project start date
   Reference to any additional project files if
    applicable.
   Team size information
Suggested Mandatory Data
   Final environmental adjustment factor
   Final scaling adjustment factor
   Hours per Months
   Product complexity measure of description
   Development tools
   Data quality qualifier (a to f) to measure the
    quality of the data.
   Development type e.g. new development or
    enhancement
Suggested Mandatory Data
   It is highly recommended to store a
    copy of the entire project type/domain,
    lifecycle and standard definition and
    estimation model. Definitions and model
    parameters change over time which
    makes it necessary to store a copy of the
    definitions along with the project which
    used them for later retrieval.
Suggested Optional Data
   Estimated total defects and if possible a breakdown
    (requirements, design, coding, documentation, bad fixes
    defects etc)
   Actual total defects and if possible a breakdown
    (requirements, design, coding, documentation, bad fixes
    defects etc)
   Environment factors list with ratings and ratings
    mapping
   Scaling factors list with ratings and ratings mapping
   Project estimated end date
   Project actual end date
Suggested Optional Data
   Effort by phase
   Total effort in person hours
   Effort by phase
   Schedule by phase, e.g. Waterfall: requirements,
    design, code, test, total.
   Number of people (and job functions) working in
    each phase (average or maximum)
   Sizing data such as FP count, Internet point,
    bottom up
   Sizing methodology parameter
Suggested Optional Data
   Volume/size information for all components,
    breakdown: new, reused, COTS and REVL.
   Programming language composition per
    component and unadjusted FP count
   Description of the programming language used
    (IDE heavy? Etc).

   Sizing methodology parameter definitions:
       Methodology used (name such as function points)
       List of factors and weights
Data for Model Calibrations
   COCOMO’s calibration routine needs the
    following inputs to calibration A and C
       Actual Effort, Schedule
       Number of projects used
       EAF (Environmental Factor overall) or all individual
        factors
       Exponent Base (driven by the model)
       Overall Scaling Factor or all individual factors
       KSLOC
       HPM
Goal/Question/Metric
   The Goal/Question/Metric (GQM) Paradigm is a
    mechanism that provides a framework for
    developing a metrics program.
   It was developed at the University of Maryland
    as a mechanism for formalizing the tasks of
    characterization, planning, construction,
    analysis, learning and feedback.
   The GQM paradigm was developed for all types
    of studies, particularly studies concerned with
    improvement issues. The paradigm does not
    provide specific goals but rather a framework for
    stating goals and refining them into questions to
    provide a specification for the data needed to
    help achieve the goals.
Goal/Question/Metric
   The GQM paradigm consists of three steps:
       1. Generate a set of goals
       2. Derive a set of questions
       3. Develop a set of metrics

   The next pages list several Goals, Questions and
    their needed metrics/data points. It can be used
    to determine what project data needs to be
    collected.
Goal/Question/Metric
   Main Goal: Improve the Software Process
       Goal: Improve Productivity
            Question: What are the overall and subgroup productivities?
                  Metric: SLOC/person-month (overall and subgroup averages).
                   Productivity should be calculated from actuals.
            Question: What is the productivity per activity?
                  Metric: SLOC/person-month per activity.
            Question: Is productivity increasing over time?
                  Metric: overall productivity trends (multiple projects).
Goal/Question/Metric
   Main Goal: Improve the Software Process
       Goal: Improve Quality
            Question: What is the overall and subgroup defect densities?
                  Metrics: Defects/KSLOC
            Question: How many defects are introduced by type?
                  Metric: Percent of defects introduced per type. A project type
                   must be specified
            Question: What defect types occur the most?
                  Metric: Defect type percentages ordered by magnitude
Goal/Question/Metric
   Main Goal: Improve software process
    predictability
       Goal: Improve Effort Predictability
            Question: How accurate are effort estimates?
                 Metric: actual vs. estimated effort

More Related Content

PPT
Estimation maturity model using function points
PPT
Managing projects by data
PPT
A Regression Analysis Approach for Building a Prediction Model for System Tes...
PDF
A defect prediction model based on the relationships between developers and c...
PPTX
Establishing A Defect Prediction Model Using A Combination of Product Metrics...
PPTX
Project Cost Management
PPTX
PMP Training - 08 project quality management
PDF
Importance of software quality metrics
Estimation maturity model using function points
Managing projects by data
A Regression Analysis Approach for Building a Prediction Model for System Tes...
A defect prediction model based on the relationships between developers and c...
Establishing A Defect Prediction Model Using A Combination of Product Metrics...
Project Cost Management
PMP Training - 08 project quality management
Importance of software quality metrics

What's hot (20)

PPSX
Project Control System
PPTX
Test planning
PDF
Planning and Optimization of Resource Constrained Project Scheduling by using...
PPTX
Project management : Project Monitoring and Control by iFour Technolab Pvt. Ltd.
DOCX
Term paper spm
PPTX
Test Plan Simplicity
PDF
Software Estimation Methodology - MVC Points
DOC
02 software test plan template
PPTX
Test Progress Monitoring and Control
PPTX
Test Planning and Test Estimation Techniques
PDF
Test plan
PPT
Monitoring&evaluation best practices
PDF
Top down
PPTX
Project monitoring lecture 1
DOC
Testplan
DOCX
EVMS for Commercial projects
PPTX
Testing Process
PPTX
Bca 5th sem seminar(software measurements)
PDF
COST CONTROL AND TRACKING OF A BUILDING BY EARNED VALUE METHOD
PPTX
Testing strategies
Project Control System
Test planning
Planning and Optimization of Resource Constrained Project Scheduling by using...
Project management : Project Monitoring and Control by iFour Technolab Pvt. Ltd.
Term paper spm
Test Plan Simplicity
Software Estimation Methodology - MVC Points
02 software test plan template
Test Progress Monitoring and Control
Test Planning and Test Estimation Techniques
Test plan
Monitoring&evaluation best practices
Top down
Project monitoring lecture 1
Testplan
EVMS for Commercial projects
Testing Process
Bca 5th sem seminar(software measurements)
COST CONTROL AND TRACKING OF A BUILDING BY EARNED VALUE METHOD
Testing strategies
Ad

Similar to Data Collection Points And Gqm (20)

PPT
Metrics Sirisha
PPT
Metrics Sirisha
PPT
Metrics Sirisha
PPT
Lecture5
PPT
Chapter 11 Metrics for process and projects.ppt
PPT
8 project planning
PDF
Using Benchmarking to Quantify the Benefits of Software Process Improvement
PPT
21UCAE52 Software Project Management.ppt
PPTX
Software Project Management Unit 2 chapters
PDF
Spm project planning
PPT
Lecture 7 Software Metrics.ppt
PPTX
Software metrics
PPTX
software metrics(process,project,product)
PPT
Lecture3
PPT
Data Collection Process And Integrity
PPTX
CS8494 SOFTWARE ENGINEERING Unit-5
PPTX
Metrics for Mofel-Based Systems Development
PPTX
Measure phase lean six sigma tollgate template
PPTX
Measure phase lean six sigma tollgate template
PPTX
Software engineering
Metrics Sirisha
Metrics Sirisha
Metrics Sirisha
Lecture5
Chapter 11 Metrics for process and projects.ppt
8 project planning
Using Benchmarking to Quantify the Benefits of Software Process Improvement
21UCAE52 Software Project Management.ppt
Software Project Management Unit 2 chapters
Spm project planning
Lecture 7 Software Metrics.ppt
Software metrics
software metrics(process,project,product)
Lecture3
Data Collection Process And Integrity
CS8494 SOFTWARE ENGINEERING Unit-5
Metrics for Mofel-Based Systems Development
Measure phase lean six sigma tollgate template
Measure phase lean six sigma tollgate template
Software engineering
Ad

Data Collection Points And Gqm

  • 1. Data Collection List of Data Points Gerrit Klaschke
  • 2. What is covered?  Suggested Mandatory Data  Optional Data  Data for Model Calibrations  Goal-Question-Metrics Approach
  • 3. Data Collection – What to Collect?  This powerpoint lists what project data needs to be collected for different purposes.  First part lists mandatory data and suggested optional data.  Second part lists data needed for model calibration  Last part: we categorize the data using the GQM approach.  After setting a goal and questions, the GQM approach defines the metrics and data points which are needed to be collected. These lists are also synchronized with sources such as ISBSG Database, USC COCOMO Project Database and QSM data collection documents available on the Internet
  • 4. Suggested Mandatory Data  Project name including version/ increment/ release if applicable  Project manager name and contact information  Company/department information  Submitter name and contact information  Submission date
  • 5. Suggested Mandatory Data  Estimation model used including the model parameter used  Project type/domain  Project lifecycle  Project standard – if applicable  Equivalent SLOC or other sizing depending on the model  Counting convention used, e.g. logical or physical SLOC and counting technique used e.g. counting tool  Estimated Effort  Estimated Schedule
  • 6. Suggested Mandatory Data  Actual Effort (person months)  Actual Schedule (months)  Project memo field for free text  Project start date  Reference to any additional project files if applicable.  Team size information
  • 7. Suggested Mandatory Data  Final environmental adjustment factor  Final scaling adjustment factor  Hours per Months  Product complexity measure of description  Development tools  Data quality qualifier (a to f) to measure the quality of the data.  Development type e.g. new development or enhancement
  • 8. Suggested Mandatory Data  It is highly recommended to store a copy of the entire project type/domain, lifecycle and standard definition and estimation model. Definitions and model parameters change over time which makes it necessary to store a copy of the definitions along with the project which used them for later retrieval.
  • 9. Suggested Optional Data  Estimated total defects and if possible a breakdown (requirements, design, coding, documentation, bad fixes defects etc)  Actual total defects and if possible a breakdown (requirements, design, coding, documentation, bad fixes defects etc)  Environment factors list with ratings and ratings mapping  Scaling factors list with ratings and ratings mapping  Project estimated end date  Project actual end date
  • 10. Suggested Optional Data  Effort by phase  Total effort in person hours  Effort by phase  Schedule by phase, e.g. Waterfall: requirements, design, code, test, total.  Number of people (and job functions) working in each phase (average or maximum)  Sizing data such as FP count, Internet point, bottom up  Sizing methodology parameter
  • 11. Suggested Optional Data  Volume/size information for all components, breakdown: new, reused, COTS and REVL.  Programming language composition per component and unadjusted FP count  Description of the programming language used (IDE heavy? Etc).  Sizing methodology parameter definitions:  Methodology used (name such as function points)  List of factors and weights
  • 12. Data for Model Calibrations  COCOMO’s calibration routine needs the following inputs to calibration A and C  Actual Effort, Schedule  Number of projects used  EAF (Environmental Factor overall) or all individual factors  Exponent Base (driven by the model)  Overall Scaling Factor or all individual factors  KSLOC  HPM
  • 13. Goal/Question/Metric  The Goal/Question/Metric (GQM) Paradigm is a mechanism that provides a framework for developing a metrics program.  It was developed at the University of Maryland as a mechanism for formalizing the tasks of characterization, planning, construction, analysis, learning and feedback.  The GQM paradigm was developed for all types of studies, particularly studies concerned with improvement issues. The paradigm does not provide specific goals but rather a framework for stating goals and refining them into questions to provide a specification for the data needed to help achieve the goals.
  • 14. Goal/Question/Metric  The GQM paradigm consists of three steps:  1. Generate a set of goals  2. Derive a set of questions  3. Develop a set of metrics  The next pages list several Goals, Questions and their needed metrics/data points. It can be used to determine what project data needs to be collected.
  • 15. Goal/Question/Metric  Main Goal: Improve the Software Process  Goal: Improve Productivity  Question: What are the overall and subgroup productivities?  Metric: SLOC/person-month (overall and subgroup averages). Productivity should be calculated from actuals.  Question: What is the productivity per activity?  Metric: SLOC/person-month per activity.  Question: Is productivity increasing over time?  Metric: overall productivity trends (multiple projects).
  • 16. Goal/Question/Metric  Main Goal: Improve the Software Process  Goal: Improve Quality  Question: What is the overall and subgroup defect densities?  Metrics: Defects/KSLOC  Question: How many defects are introduced by type?  Metric: Percent of defects introduced per type. A project type must be specified  Question: What defect types occur the most?  Metric: Defect type percentages ordered by magnitude
  • 17. Goal/Question/Metric  Main Goal: Improve software process predictability  Goal: Improve Effort Predictability  Question: How accurate are effort estimates?  Metric: actual vs. estimated effort