Good Machine Learning Practice
for Medical Device Development
Guiding principles
Contact Digitalhealth@fda.hhs.gov
Multi-Disciplinary Expertise Is Leveraged
Throughout the Total Product Life Cycle
In-depth understanding of a model’s intended integration into clinical
workflow, and the desired benefits and associated patient risks, can help
ensure that ML-enabled medical devices are safe and effective and address
clinically meaningful needs over the lifecycle of the device.
Good Software Engineering and Security
Practices Are Implemented
Model design is implemented with attention to the “fundamentals”: good
software engineering practices, data quality assurance, data management,
and robust cybersecurity practices. These practices include methodical risk
management and design process that can appropriately capture and
communicate design, implementation, and risk management decisions and
rationale, as well as ensure data authenticity and integrity.
Clinical Study Participants and Data Sets
Are Representative of the Intended Patient
Population
Data collection protocols should ensure that the relevant characteristics of the
intended patient population (for example, in terms of age, gender, sex, race,
and ethnicity), use, and measurement inputs are sufficiently represented in a
sample of adequate size in the clinical study and training and test datasets, so
that results can be reasonably generalized to the population of interest. This is
important to manage any bias, promote appropriate and generalizable
performance across the intended patient population, assess usability, and
identify circumstances where the model may underperform.
Training Data Sets Are
Independent of Test Sets
Training and test datasets are selected and maintained to be appropriately
independent of one another. All potential sources of dependence, including
patient, data acquisition, and site factors, are considered and addressed to
assure independence.
Selected Reference Datasets Are Based
Upon Best Available Methods
Accepted, best available methods for developing a reference dataset (that is, a
reference standard) ensure that clinically relevant and well characterized data
are collected and the limitations of the reference are understood. If available,
accepted reference datasets in model development and testing that promote
and demonstrate model robustness and generalizability across the intended
patient population are used.
Model Design Is Tailored to the Available
Data and Reflects the Intended Use of the
Device
Model design is suited to the available data and supports the active mitigation
of known risks, like overfitting, performance degradation, and security risks.
The clinical benefits and risks related to the product are well understood,
used to derive clinically meaningful performance goals for testing, and
support that the product can safely and effectively achieve its intended use.
Considerations include the impact of both global and local performance and
uncertainty/variability in the device inputs, outputs, intended patient
populations, and clinical use conditions.
Focus Is Placed on the Performance of the
Human-AI Team
Where the model has a “human in the loop,” human factors considerations
and the human interpretability of the model outputs are addressed with
emphasis on the performance of the Human-AI team, rather than just the
performance of the model in isolation.
Testing Demonstrates Device Performance
during Clinically Relevant Conditions
Statistically sound test plans are developed and executed to generate clinically
relevant device performance information independently of the training data
set. Considerations include the intended patient population, important
subgroups, clinical environment and use by the Human-AI team,
measurement inputs, and potential confounding factors.
Users Are Provided Clear,
Essential Information
Users are provided ready access to clear, contextually relevant information
that is appropriate for the intended audience (such as health care providers or
patients) including: the product’s intended use and indications for use,
performance of the model for appropriate subgroups, characteristics of the
data used to train and test the model, acceptable inputs, known limitations,
user interface interpretation, and clinical workflow integration of the model.
Users are also made aware of device modifications and updates from real-
world performance monitoring, the basis for decision-making when available,
and a means to communicate product concerns to the developer.
Deployed Models Are Monitored for
Performance and Re-training Risks are
Managed
Deployed models have the capability to be monitored in “real world” use with
a focus on maintained or improved safety and performance. Additionally,
when models are periodically or continually trained after deployment, there
are appropriate controls in place to manage risks of overfitting, unintended
bias, or degradation of the model (for example, dataset drift) that may impact
the safety and performance of the model as it is used by the Human-AI team.

More Related Content

PPTX
Research Data Management for Clinical Trials and Quality Improvement
PPTX
Rules for Healthcare Transformation Frameworks and Tools v1
PPTX
lecture 9 B
PDF
Patient reported outcomes
DOCX
Anusha Kusuma_PV
PPTX
Case report form and application
PDF
JonellAgyeiSakyiResume
PPTX
RISK BASED MONITORING
Research Data Management for Clinical Trials and Quality Improvement
Rules for Healthcare Transformation Frameworks and Tools v1
lecture 9 B
Patient reported outcomes
Anusha Kusuma_PV
Case report form and application
JonellAgyeiSakyiResume
RISK BASED MONITORING

What's hot (20)

PPT
Feb 2009: my University of Wisconsin colloquim presentation
PPT
Elmallah june27 11am_room230_a
PDF
eMagzine Spring 2016 (Trusha Hake)
PPT
Jason Saleem Slides from AHRQ Kick-Off
PPTX
Lecture 9 A
PDF
Knowledge Translation Tool Dr. I Hassan V14
PPT
Brad Doebbeling Slides for AHRQ Kick-Off Event
PPT
Heather Woodward Slides from AHRQ Kick-Off
PPT
David Haggstrom Slides from AHRQ Kick-Off Event
PDF
ESSA_2016_FC_Poster_FINAL
PDF
Integrating Trial Data Processes Across Functional Areas using Electronic Dat...
PPTX
Medical Simulation 3.0
PPTX
Patient Record System
PDF
Defining a Central Monitoring Capability: Sharing the Experience of TransCele...
PDF
Risk Based Monitoring in Practice
PDF
Technology Considerations to Enable the Risk-Based Monitoring Methodology
PPTX
Week5hcs451 presentation1
PPT
Ingenix Hc Conference Pres 042108
PPTX
Ensuring high quality data
PPTX
Network integration
Feb 2009: my University of Wisconsin colloquim presentation
Elmallah june27 11am_room230_a
eMagzine Spring 2016 (Trusha Hake)
Jason Saleem Slides from AHRQ Kick-Off
Lecture 9 A
Knowledge Translation Tool Dr. I Hassan V14
Brad Doebbeling Slides for AHRQ Kick-Off Event
Heather Woodward Slides from AHRQ Kick-Off
David Haggstrom Slides from AHRQ Kick-Off Event
ESSA_2016_FC_Poster_FINAL
Integrating Trial Data Processes Across Functional Areas using Electronic Dat...
Medical Simulation 3.0
Patient Record System
Defining a Central Monitoring Capability: Sharing the Experience of TransCele...
Risk Based Monitoring in Practice
Technology Considerations to Enable the Risk-Based Monitoring Methodology
Week5hcs451 presentation1
Ingenix Hc Conference Pres 042108
Ensuring high quality data
Network integration
Ad

Similar to FDA Good Machine Learning Practices (20)

PDF
Unleashing the Power of Data: Enhancing Physician Outreach through Machine Le...
PPTX
mICF research
PPTX
mICF poster 5 (olaf) micf research
PPT
Team Sol2 01 Health Care Informatics Power Point
PDF
The Key Components of a Successful Clinical Evaluation Report under MDR 2017 ...
PDF
PSQH July-Aug 2015 Simplified ST Model - Woods-Pestotnik
PPTX
Ulcer Prediction in this world of AI.pptx
PDF
PPTX
HXR 2017: Bakul Patel: How the FDA Is Promoting Innovation and Protecting the...
PDF
NURS FPX 6410 Assessment 3 Exploration of Regulations and Implications for Pr...
PPTX
Computer Information Systems and the Electronic Health Record
DOCX
Quality management model
DOCX
Quality management system model
DOCX
STUDY PROTOCOL Open AccessSafety Assurance Factors for Ele.docx
PDF
Exploring-Human-Factors-and-Usability-in-Medical-Devices.pdf
DOCX
Out-Patient Physical therapy risk managementGrand Canyon Unive
PDF
BME 307 - HMIS - Data Management Systems 24112021 Final.pdf
PDF
Big data, RWE and AI in Clinical Trials made simple
DOCX
Quality risk management sop
PPTX
Patients outcomes
Unleashing the Power of Data: Enhancing Physician Outreach through Machine Le...
mICF research
mICF poster 5 (olaf) micf research
Team Sol2 01 Health Care Informatics Power Point
The Key Components of a Successful Clinical Evaluation Report under MDR 2017 ...
PSQH July-Aug 2015 Simplified ST Model - Woods-Pestotnik
Ulcer Prediction in this world of AI.pptx
HXR 2017: Bakul Patel: How the FDA Is Promoting Innovation and Protecting the...
NURS FPX 6410 Assessment 3 Exploration of Regulations and Implications for Pr...
Computer Information Systems and the Electronic Health Record
Quality management model
Quality management system model
STUDY PROTOCOL Open AccessSafety Assurance Factors for Ele.docx
Exploring-Human-Factors-and-Usability-in-Medical-Devices.pdf
Out-Patient Physical therapy risk managementGrand Canyon Unive
BME 307 - HMIS - Data Management Systems 24112021 Final.pdf
Big data, RWE and AI in Clinical Trials made simple
Quality risk management sop
Patients outcomes
Ad

More from Arlen Meyers, MD, MBA (20)

PPTX
How to manage technology risk and corporate growth
PPTX
Are your ready to innovate: The physician entrepreneurs checklist
PPTX
Decision making like a poker player would
PPTX
Entrepreneur Personas: Combinations of willing and able
PPTX
Bioengineering Technopreneur Tips: Are you ready to innovate?
PPTX
We need more entrepreneurial and intrapreneurial academic medical centers
PPTX
How to build a successful medical tourism practice
PPTX
Are you ready to innovate: Means, mindset, motivation?
PPTX
3 steps to AI intrapreneurship: How to get your in-house developed AI product...
PPTX
How to be a successful physician intrapreneur
PPTX
Mission critical AMC entrepreneurship.pptx
PPTX
How to close CS/engineering doctor divide.pptx
PPTX
Entrepreneur Personas.pptx
PPTX
Sope School Intro to AI entrepreneurship.pptx
PPTX
MedEd 2030.pptx
PPTX
What sick care technology aintrepreneurs should know.pptx
PPTX
Ten things to know about bioentrepreneurship
PPTX
Healthcare AI competencies.pptx
PPTX
PPTX
Sope School Intro to AI entrepreneurship.pptx
How to manage technology risk and corporate growth
Are your ready to innovate: The physician entrepreneurs checklist
Decision making like a poker player would
Entrepreneur Personas: Combinations of willing and able
Bioengineering Technopreneur Tips: Are you ready to innovate?
We need more entrepreneurial and intrapreneurial academic medical centers
How to build a successful medical tourism practice
Are you ready to innovate: Means, mindset, motivation?
3 steps to AI intrapreneurship: How to get your in-house developed AI product...
How to be a successful physician intrapreneur
Mission critical AMC entrepreneurship.pptx
How to close CS/engineering doctor divide.pptx
Entrepreneur Personas.pptx
Sope School Intro to AI entrepreneurship.pptx
MedEd 2030.pptx
What sick care technology aintrepreneurs should know.pptx
Ten things to know about bioentrepreneurship
Healthcare AI competencies.pptx
Sope School Intro to AI entrepreneurship.pptx

Recently uploaded (20)

PDF
20250617 - IR - Global Guide for HR - 51 pages.pdf
PPTX
Software Engineering and software moduleing
PPTX
ASME PCC-02 TRAINING -DESKTOP-NLE5HNP.pptx
PPTX
CN_Unite_1 AI&DS ENGGERING SPPU PUNE UNIVERSITY
PDF
Exploratory_Data_Analysis_Fundamentals.pdf
PPTX
Graph Data Structures with Types, Traversals, Connectivity, and Real-Life App...
PPTX
mechattonicsand iotwith sensor and actuator
PPTX
CyberSecurity Mobile and Wireless Devices
PPTX
Module 8- Technological and Communication Skills.pptx
PDF
LOW POWER CLASS AB SI POWER AMPLIFIER FOR WIRELESS MEDICAL SENSOR NETWORK
PPTX
AUTOMOTIVE ENGINE MANAGEMENT (MECHATRONICS).pptx
PDF
distributed database system" (DDBS) is often used to refer to both the distri...
PDF
MLpara ingenieira CIVIL, meca Y AMBIENTAL
PPTX
Chemical Technological Processes, Feasibility Study and Chemical Process Indu...
PDF
Soil Improvement Techniques Note - Rabbi
PPTX
wireless networks, mobile computing.pptx
PDF
Artificial Superintelligence (ASI) Alliance Vision Paper.pdf
PPTX
ai_satellite_crop_management_20250815030350.pptx
PDF
Computer organization and architecuture Digital Notes....pdf
PPTX
Amdahl’s law is explained in the above power point presentations
20250617 - IR - Global Guide for HR - 51 pages.pdf
Software Engineering and software moduleing
ASME PCC-02 TRAINING -DESKTOP-NLE5HNP.pptx
CN_Unite_1 AI&DS ENGGERING SPPU PUNE UNIVERSITY
Exploratory_Data_Analysis_Fundamentals.pdf
Graph Data Structures with Types, Traversals, Connectivity, and Real-Life App...
mechattonicsand iotwith sensor and actuator
CyberSecurity Mobile and Wireless Devices
Module 8- Technological and Communication Skills.pptx
LOW POWER CLASS AB SI POWER AMPLIFIER FOR WIRELESS MEDICAL SENSOR NETWORK
AUTOMOTIVE ENGINE MANAGEMENT (MECHATRONICS).pptx
distributed database system" (DDBS) is often used to refer to both the distri...
MLpara ingenieira CIVIL, meca Y AMBIENTAL
Chemical Technological Processes, Feasibility Study and Chemical Process Indu...
Soil Improvement Techniques Note - Rabbi
wireless networks, mobile computing.pptx
Artificial Superintelligence (ASI) Alliance Vision Paper.pdf
ai_satellite_crop_management_20250815030350.pptx
Computer organization and architecuture Digital Notes....pdf
Amdahl’s law is explained in the above power point presentations

FDA Good Machine Learning Practices

  • 1. Good Machine Learning Practice for Medical Device Development Guiding principles Contact Digitalhealth@fda.hhs.gov
  • 2. Multi-Disciplinary Expertise Is Leveraged Throughout the Total Product Life Cycle In-depth understanding of a model’s intended integration into clinical workflow, and the desired benefits and associated patient risks, can help ensure that ML-enabled medical devices are safe and effective and address clinically meaningful needs over the lifecycle of the device.
  • 3. Good Software Engineering and Security Practices Are Implemented Model design is implemented with attention to the “fundamentals”: good software engineering practices, data quality assurance, data management, and robust cybersecurity practices. These practices include methodical risk management and design process that can appropriately capture and communicate design, implementation, and risk management decisions and rationale, as well as ensure data authenticity and integrity.
  • 4. Clinical Study Participants and Data Sets Are Representative of the Intended Patient Population Data collection protocols should ensure that the relevant characteristics of the intended patient population (for example, in terms of age, gender, sex, race, and ethnicity), use, and measurement inputs are sufficiently represented in a sample of adequate size in the clinical study and training and test datasets, so that results can be reasonably generalized to the population of interest. This is important to manage any bias, promote appropriate and generalizable performance across the intended patient population, assess usability, and identify circumstances where the model may underperform.
  • 5. Training Data Sets Are Independent of Test Sets Training and test datasets are selected and maintained to be appropriately independent of one another. All potential sources of dependence, including patient, data acquisition, and site factors, are considered and addressed to assure independence.
  • 6. Selected Reference Datasets Are Based Upon Best Available Methods Accepted, best available methods for developing a reference dataset (that is, a reference standard) ensure that clinically relevant and well characterized data are collected and the limitations of the reference are understood. If available, accepted reference datasets in model development and testing that promote and demonstrate model robustness and generalizability across the intended patient population are used.
  • 7. Model Design Is Tailored to the Available Data and Reflects the Intended Use of the Device Model design is suited to the available data and supports the active mitigation of known risks, like overfitting, performance degradation, and security risks. The clinical benefits and risks related to the product are well understood, used to derive clinically meaningful performance goals for testing, and support that the product can safely and effectively achieve its intended use. Considerations include the impact of both global and local performance and uncertainty/variability in the device inputs, outputs, intended patient populations, and clinical use conditions.
  • 8. Focus Is Placed on the Performance of the Human-AI Team Where the model has a “human in the loop,” human factors considerations and the human interpretability of the model outputs are addressed with emphasis on the performance of the Human-AI team, rather than just the performance of the model in isolation.
  • 9. Testing Demonstrates Device Performance during Clinically Relevant Conditions Statistically sound test plans are developed and executed to generate clinically relevant device performance information independently of the training data set. Considerations include the intended patient population, important subgroups, clinical environment and use by the Human-AI team, measurement inputs, and potential confounding factors.
  • 10. Users Are Provided Clear, Essential Information Users are provided ready access to clear, contextually relevant information that is appropriate for the intended audience (such as health care providers or patients) including: the product’s intended use and indications for use, performance of the model for appropriate subgroups, characteristics of the data used to train and test the model, acceptable inputs, known limitations, user interface interpretation, and clinical workflow integration of the model. Users are also made aware of device modifications and updates from real- world performance monitoring, the basis for decision-making when available, and a means to communicate product concerns to the developer.
  • 11. Deployed Models Are Monitored for Performance and Re-training Risks are Managed Deployed models have the capability to be monitored in “real world” use with a focus on maintained or improved safety and performance. Additionally, when models are periodically or continually trained after deployment, there are appropriate controls in place to manage risks of overfitting, unintended bias, or degradation of the model (for example, dataset drift) that may impact the safety and performance of the model as it is used by the Human-AI team.