Assyro AI
Assyro AI logo background
ai ml medical device fda
predetermined change control plan
ai medical device regulation
fda ai ml action plan

AI/ML Medical Device FDA Guidance: Predetermined Change Control Plans

Guide

AI/ML medical device FDA guidance covering predetermined change control plans, total product lifecycle approach, GMLP principles, and real-world monitoring.

Assyro Team
16 min read

AI/ML Medical Device FDA Guidance: Predetermined Change Control Plans

Quick Answer

FDA regulates AI-enabled medical devices through the existing device framework for software functions, including 510(k), De Novo, and PMA pathways where applicable. FDA's August 2025 final guidance on predetermined change control plans (PCCPs) explains how a manufacturer can describe certain planned AI-enabled device modifications in a marketing submission so those changes can later be implemented within an authorized PCCP. FDA also publishes an AI/ML-enabled medical device list and recommends a total product lifecycle approach that combines premarket review, postmarket monitoring, and controlled change management.

Key Takeaways

Key Takeaways

  • FDA's public framework for AI-enabled devices is built around manufacturer-controlled changes, including changes managed through a reviewed PCCP.
  • PCCPs must include three elements: specific modification description, impact assessment, and description of the modified device with quantitative acceptance criteria.
  • The ten GMLP principles (FDA/HC/MHRA, October 2021) are not legally binding but FDA expects AI/ML submissions to address them, particularly representative data, independent datasets, and post-deployment monitoring.
  • FDA maintains a public AI/ML-enabled device list, but sponsors should rely on the FDA list itself rather than informal market-size counts.

FDA's AI/ML Regulatory Evolution

FDA's approach to AI/ML medical devices has evolved through a series of discussion papers, action plans, and guidance documents:

Timeline of Key FDA Actions

DateActionSignificance
April 2019Proposed Regulatory Framework for Modifications to AI/ML-Based SaMD (Discussion Paper)First comprehensive articulation of FDA's vision for AI/ML device regulation
January 2021AI/ML-Based SaMD Action PlanFive-point action plan responding to public comments on the 2019 discussion paper
September 2021Draft Guidance: PCCP for AI/ML-Enabled Device Software FunctionsFirst draft of PCCP guidance
October 2021Good Machine Learning Practice (GMLP) Guiding Principles (with HC, MHRA)Ten principles for responsible AI/ML development
August 2025Final Guidance: Marketing Submission Recommendations for a PCCP for AI-Enabled Device Software FunctionsFinalized PCCP framework
June 2023Content of Premarket Submissions for Device Software Functions (updated)Updated software documentation guidance applicable to AI/ML
October 2023Draft Guidance: Lifecycle Management Considerations for AI/ML-Enabled Device Software FunctionsProposed framework for ongoing AI/ML device management
September 2024Draft Guidance: AI-Enabled Device Software Functions: Lifecycle Management and Marketing Submission RecommendationsUpdated and expanded lifecycle guidance

Locked vs Adaptive AI/ML Algorithms

A fundamental distinction in AI/ML device regulation is between locked and adaptive algorithms.

Locked Algorithms

A locked algorithm is one whose function does not change after deployment. The algorithm produces the same output for a given input every time, and the manufacturer does not update the model in response to new data during clinical use. The algorithm is fixed at the time of premarket authorization.

Locked algorithms include:

  • Rule-based systems with fixed decision logic
  • Machine learning models trained on a fixed dataset and deployed without retraining
  • Neural networks with frozen weights

Locked algorithms are regulated through traditional premarket review. Any change to the algorithm (retraining, weight updates, rule modifications) requires either a new premarket submission or coverage under an authorized PCCP.

Adaptive Algorithms

An adaptive algorithm changes its behavior over time based on new data or learning. The algorithm's output for a given input may change as it processes additional data. Adaptive algorithms include:

  • Continuously learning algorithms that update model parameters based on new patient data
  • Algorithms that periodically retrain on expanded datasets
  • Reinforcement learning systems that modify behavior based on outcome feedback

FDA's 2019 discussion paper distinguished between:

  • Algorithms that adapt based on data from the local institution: Higher risk because changes are site-specific and may not be validated across diverse populations
  • Algorithms that adapt based on centralized data aggregation: Modifications are developed centrally by the manufacturer and deployed as updates

Adaptive algorithms present regulatory challenges because the device that was authorized at the time of premarket review may behave differently in clinical use. FDA's PCCP framework is designed to address this by requiring manufacturers to pre-specify the types of adaptations, validation methodology, and performance bounds.

Current Regulatory Status

FDA's current public guidance is written around premarket-reviewed software functions and manufacturer-controlled post-authorization modifications. Sponsors developing software that could change after authorization should expect FDA to focus on clearly bounded modifications, validation methods, and postmarket monitoring rather than open-ended field learning.

Predetermined Change Control Plans in Detail

FDA's August 2025 final guidance, "Marketing Submission Recommendations for a Predetermined Change Control Plan for Artificial Intelligence-Enabled Device Software Functions," establishes the current framework for PCCPs.

PCCP Components

A PCCP must include three elements:

1. Description of Modifications (Modification Protocol)

The PCCP must describe:

  • What will change: Specific aspects of the AI/ML algorithm that may be modified (e.g., model architecture, training data, feature engineering, decision thresholds)
  • Why it will change: The rationale for the planned modification (e.g., performance improvement, expanded clinical scope, new device platform)
  • How it will change: The methodology for implementing the modification (e.g., retraining protocol, data selection criteria, validation procedures)
  • Bounds of the change: Quantitative or qualitative limits on the extent of modification

The modification description must be specific enough that FDA can evaluate the risk profile of the proposed changes. Vague descriptions such as "the algorithm may be updated to improve performance" are insufficient.

2. Impact Assessment

The PCCP must include an assessment of how the planned modifications will affect:

  • Safety: Analysis of potential risks introduced by the modification, including new failure modes
  • Effectiveness: Analysis of how the modification may affect device performance, including the possibility of performance degradation
  • Benefits: Expected improvements in device performance or clinical utility

The impact assessment should reference the device's risk analysis (per ISO 14971) and demonstrate that the manufacturer has considered the full range of potential impacts.

3. Description of Modified Device

The PCCP must describe:

  • Performance specifications: Quantitative performance metrics that the modified device must meet (e.g., sensitivity, specificity, AUC, positive predictive value)
  • Acceptance criteria: Criteria that determine whether a modification passes validation and can be deployed
  • Labeling updates: Any changes to labeling that will result from the modification

PCCP Evaluation Criteria

FDA evaluates PCCPs against the following criteria:

CriterionWhat FDA Assesses
SpecificityAre the planned modifications described with sufficient detail?
Bounded scopeAre the modifications limited to a defined scope?
Validation methodologyIs the validation approach appropriate for the type of modification?
Performance metricsAre acceptance criteria quantitative, pre-specified, and clinically meaningful?
Risk analysisHas the manufacturer identified and mitigated risks from modifications?
TransparencyWill users be informed of modifications?
MonitoringWill post-deployment performance be monitored?

PCCP Examples

Example 1: Retraining with Expanded Dataset

A manufacturer of an AI-based chest X-ray CADe system might include a PCCP that allows:

  • Retraining the model on additional chest X-ray datasets meeting pre-specified diversity and quality criteria
  • Maintaining pre-specified clinical performance criteria on a standardized test set
  • Validating performance on a held-out dataset with pre-specified demographic subgroups
  • Deploying the updated model only if all acceptance criteria are met

Example 2: Decision Threshold Adjustment

A manufacturer of an AI-based diabetic retinopathy screening device might include a PCCP that allows:

  • Adjusting the binary classification threshold within a pre-specified range (e.g., 0.40-0.60)
  • Validating each threshold change on a standardized test set with known clinical outcomes
  • Maintaining pre-specified clinical performance criteria at any selected threshold

Example 3: Platform Expansion

A manufacturer might include a PCCP that allows:

  • Deploying the same algorithm on additional hardware platforms (e.g., new tablet models, new operating system versions)
  • Validating that algorithm performance is equivalent across platforms
  • Documenting any platform-specific adjustments to image preprocessing

What a PCCP Cannot Cover

FDA's PCCP guidance clarifies that certain changes fall outside the scope of a PCCP:

  • Changes to the intended use or indications for use
  • Changes to the target population (e.g., expanding from adult to pediatric use)
  • Changes to the clinical domain (e.g., expanding from chest X-ray to brain MRI)
  • Changes that would require a different classification or regulatory pathway
  • Open-ended modifications without specific bounds

These changes require a new premarket submission (510(k), De Novo, or PMA supplement).

Total Product Lifecycle Approach

FDA's TPLC approach for AI/ML devices envisions a continuous cycle of oversight rather than a one-time premarket review.

TPLC Framework

[@portabletext/react] Unknown block type "code", specify a component for it in the `components.types` prop

Key TPLC Elements

1. Premarket Authorization:

  • Traditional premarket review (510(k), De Novo, PMA) with PCCP included
  • Clinical evidence supporting initial algorithm performance
  • Software documentation per FDA guidance

2. Post-Market Monitoring:

  • Real-world performance monitoring comparing actual performance to premarket testing
  • Adverse event reporting under 21 CFR Part 803
  • Any additional periodic reporting should follow the reporting structure FDA describes in the applicable authorization, guidance, or future program-specific communications

3. Modification Management:

  • Modifications within PCCP scope: implement per PCCP methodology, document, report to FDA
  • Modifications outside PCCP scope: new premarket submission required
  • PCCP updates: manufacturer may submit updated PCCPs through supplemental submissions

4. Transparency:

  • Labeling that describes the AI/ML nature of the device
  • Disclosure of algorithm modifications to users
  • Public listing of AI/ML-enabled devices on FDA's AI/ML database

Good Machine Learning Practice (GMLP)

In October 2021, FDA, Health Canada, and the UK's MHRA jointly published "Good Machine Learning Practice for Medical Device Development: Guiding Principles." These ten principles provide a framework for responsible AI/ML device development:

The Ten GMLP Principles

#PrincipleCore Requirement
1Multi-disciplinary expertiseDevelopment team includes clinical, technical, regulatory, and human factors expertise
2Good software engineering and security practicesSoftware development follows established best practices (e.g., IEC 62304)
3Representative clinical study participants and datasetsTraining and testing data represent the intended patient population
4Independent training, tuning, and test datasetsSeparation of data used for training, hyperparameter tuning, and testing
5Reference datasets based on best available methodsGround truth labels established through clinically appropriate methods
6Tailored model design to intended useModel complexity appropriate for the clinical task and data
7Focus on performance of human-AI teamEvaluation considers how clinicians interact with AI output
8Clinical study testing demonstrates device performance in clinically relevant conditionsTesting reflects real-world clinical deployment, not just curated datasets
9Users provided clear, essential informationLabeling includes description of AI/ML function, limitations, and intended use
10Deployed models monitored and managed for re-training risksPost-deployment performance monitoring with procedures for addressing degradation

GMLP and Regulatory Submissions

While the GMLP principles are not legally binding requirements, FDA expects AI/ML device submissions to address these principles. Specifically:

  • Principle 3 (Representative data): FDA routinely requests demographic breakdowns of training and testing data and evaluates performance across subgroups
  • Principle 4 (Independent datasets): FDA expects clear documentation of dataset partitioning and absence of data leakage
  • Principle 7 (Human-AI team performance): For CADe/CADx devices, FDA typically requires reader studies evaluating clinician performance with and without the AI
  • Principle 10 (Post-deployment monitoring): FDA may impose post-market surveillance requirements, particularly through Special Controls in De Novo orders

Transparency and Labeling

FDA's Transparency Expectations

FDA expects AI/ML device labeling to include:

Labeling ElementContent
Device descriptionExplanation of the AI/ML function, including general description of the algorithm type
Intended useSpecific clinical use case, target population, use environment
Inputs and outputsWhat data the algorithm processes and what output it provides
Performance dataSummary of clinical performance (sensitivity, specificity, AUC, etc.) with confidence intervals
Subgroup performancePerformance broken down by relevant demographic and clinical subgroups
LimitationsKnown limitations, populations not studied, conditions under which performance may degrade
WarningsSituations where the AI output should not be relied upon
Training data descriptionGeneral description of the training data characteristics (size, demographics, source, labeling method)

FDA's AI/ML-Enabled Device Database

FDA maintains a publicly accessible list of AI/ML-enabled devices that have received marketing authorization. Sponsors should use that official FDA list for current product counts, specialty distribution, and authorization status rather than relying on secondary summaries.

The concentration in radiology reflects both the maturity of AI/ML applications in medical imaging and the availability of large, labeled imaging datasets for algorithm training.

Real-World Performance Monitoring

Why Post-Market Monitoring Matters for AI/ML

AI/ML algorithms can experience performance degradation in real-world deployment due to:

  • Dataset shift: Differences between the training data distribution and the real-world data distribution
  • Covariate shift: Changes in patient demographics, disease prevalence, or clinical practice patterns
  • Concept drift: Changes in the relationship between input features and the target outcome
  • Acquisition shift: Changes in data acquisition equipment, protocols, or settings (particularly relevant for medical imaging)

FDA's Evolving Monitoring Framework

FDA's October 2023 draft guidance "Lifecycle Management Considerations for AI/ML-Enabled Device Software Functions" proposed a framework for ongoing performance monitoring, including:

Performance Metrics:

  • Tracking key performance metrics (sensitivity, specificity, PPV, NPV) against pre-specified thresholds
  • Monitoring for algorithmic bias through subgroup performance analysis
  • Detecting performance degradation through statistical process control methods

Reporting Requirements:

  • Adverse event reporting per 21 CFR Part 803 (mandatory)
  • Periodic performance reports (proposed; format and frequency under development)
  • Notification to FDA of performance degradation below pre-specified thresholds

Corrective Actions:

  • Retraining or updating the algorithm within PCCP scope
  • Issuing field safety corrective actions if performance falls below safety thresholds
  • Recalling or withdrawing the device if safety issues cannot be resolved

Bias and Fairness

FDA has identified algorithmic bias as a significant concern for AI/ML medical devices. Sources of bias include:

Bias SourceDescriptionMitigation
Training data biasUnderrepresentation of demographic groups in training dataEnsure training data diversity, oversample underrepresented groups
Label biasSystematic errors in ground truth labels correlated with demographicsUse multiple labelers, audit label quality across subgroups
Selection biasNon-random patient selection in training dataDocument selection criteria, evaluate generalizability
Measurement biasSystematic differences in data acquisition across patient populationsStandardize acquisition protocols, test across equipment
Aggregation biasSingle model applied to distinct clinical subgroups without accounting for group-specific patternsEvaluate performance in relevant subgroups, consider subgroup-specific models

FDA expects premarket submissions to address bias by:

  • Describing the demographic composition of training and testing datasets
  • Reporting performance metrics stratified by relevant demographic subgroups (age, sex, race/ethnicity, disease severity)
  • Identifying populations where the device has not been evaluated
  • Describing the manufacturer's approach to bias detection and mitigation

Practical Implications for AI/ML Device Development

Pre-Submission Strategy

AI/ML device manufacturers should strongly consider a Pre-Submission (Q-Sub) meeting with FDA to discuss:

  1. Whether the device meets the SaMD definition
  2. Classification and regulatory pathway
  3. Clinical evidence expectations, including study design and endpoints
  4. PCCP scope and content
  5. Algorithm documentation requirements
  6. Cybersecurity considerations
  7. Post-market monitoring expectations

Documentation Best Practices

Documentation AreaRecommended Content
Algorithm descriptionArchitecture, input/output specification, preprocessing steps, model type, training procedure
Training dataSource, size, demographics, labeling methodology, quality assurance procedures
Dataset partitioningTraining/validation/test split methodology, absence of data leakage
Performance testingTest set composition, performance metrics with confidence intervals, subgroup analysis
Human factorsIntended use environment, user profile, output presentation, user study results
PCCP (if applicable)Modification protocol, impact assessment, acceptance criteria, monitoring plan
Risk analysisPer ISO 14971, including AI/ML-specific failure modes

Common FDA Review Issues

Based on publicly available FDA review summaries and decision documents, common issues in AI/ML device submissions include:

  1. Insufficient subgroup analysis: Performance not broken down by relevant demographics
  2. Training data not representative: Training data does not reflect the intended patient population
  3. Unclear algorithm description: Insufficient detail about the AI/ML methodology
  4. Missing standalone performance data: Only reader study data provided without standalone algorithm performance
  5. Inadequate ground truth methodology: Reference standard for labeling not clearly defined or clinically appropriate
  6. Cybersecurity gaps: Missing SBOM, inadequate vulnerability management plan
  7. PCCP too broad: Planned modifications described without sufficient specificity or bounded acceptance criteria

Key Regulatory References

DocumentSourceYear
Proposed Regulatory Framework for Modifications to AI/ML-Based SaMDFDA2019
AI/ML-Based SaMD Action PlanFDA2021
Good Machine Learning Practice Guiding PrinciplesFDA/HC/MHRA2021
Marketing Submission Recommendations for a PCCP for AI-Enabled Device Software FunctionsFDA2025 (final)
Content of Premarket Submissions for Device Software FunctionsFDA2023
Cybersecurity in Medical DevicesFDA2023
Clinical Performance Assessment: Considerations for CAD Devices Applied to RadiologyFDA2020
Lifecycle Management Considerations for AI/ML-Enabled Device Software FunctionsFDA2023 (draft)
AI-Enabled Device Software Functions: Lifecycle Management and Marketing Submission RecommendationsFDA2024 (draft)
AI/ML-Enabled Medical Device ListFDAContinuously updated

References