FDA regulatory compliance and medical AI certification
Resource Guide

FDA AI Training Data Requirements

What medical device developers need to know about FDA training data quality expectations, from data collection through regulatory submission.

FDA Framework Overview

The FDA regulates AI/ML-based medical devices through existing regulatory pathways (510(k), De Novo, and PMA), supplemented by AI-specific guidance documents. As of 2025, the FDA has authorized over 950 AI/ML-enabled medical devices, spanning radiology, cardiology, ophthalmology, pathology, and other specialties.

Key guidance documents that affect training data requirements include the Artificial Intelligence and Machine Learning (AI/ML) Software as a Medical Device Action Plan (2021), Good Machine Learning Practice (GMLP) guiding principles (2021, co-published with Health Canada and MHRA), and the Predetermined Change Control Plan guidance (finalized 2024).

While the FDA does not prescribe specific technical implementations for training data, their guidance establishes clear expectations around data quality, diversity, documentation, and annotator qualifications. Understanding these expectations early in development prevents costly rework during the submission process.

Data Quality Requirements

The FDA evaluates training data quality across several dimensions. Each dimension contributes to the overall confidence that your AI model will perform safely and effectively in clinical practice.

Relevance

Training data must be relevant to the device's intended use, target population, and clinical setting. A chest X-ray AI trained only on adult data cannot claim performance in pediatric populations. Data should reflect the clinical conditions, image acquisition protocols, and patient demographics the device will encounter in real-world deployment.

Representativeness

Datasets must adequately represent the intended patient population across demographics (age, sex, race/ethnicity), disease prevalence, disease severity spectrum, imaging equipment manufacturers, and clinical sites. The FDA specifically looks for evidence that datasets are not biased toward specific subgroups.

Reference Standard Quality

Labels used as ground truth must be established using an appropriate reference standard, typically expert clinician annotation, biopsy-confirmed pathology, or clinical follow-up. The FDA expects documentation of who created labels, their qualifications, the methodology used, and inter-annotator agreement metrics.

Independence

Training, validation, and test sets must be independent with no data leakage between splits. The FDA examines splitting methodology to ensure that images from the same patient do not appear in both training and test sets, a common error that artificially inflates reported performance.

Documentation Standards

FDA submissions for AI/ML devices require detailed training data documentation. The following elements are expected in your performance assessment:

Data Source Description

Document where data was collected, from how many sites, over what time period, using which equipment models and acquisition protocols. Include IRB or ethics committee approval documentation. Multi-site data collection strengthens generalizability claims.

Inclusion and Exclusion Criteria

Define which cases were included in the dataset and which were excluded, with justification. Exclusion criteria should not inadvertently remove difficult cases that the device will encounter in deployment. Document how exclusion decisions were made and by whom.

Annotation Methodology

Describe the annotation schema, tools used, annotator training process, quality control mechanisms, and adjudication procedures for disagreements. Include annotator credential documentation: board certifications, specialty training, years of clinical experience.

Quality Metrics

Report inter-annotator agreement using appropriate metrics (Cohen's Kappa for classification, Dice coefficient for segmentation). Document how disagreements were resolved and the percentage of cases requiring adjudication. Provide accuracy benchmarks against established reference standards.

Demographic Analysis

Provide breakdown of dataset demographics (age distribution, sex ratio, race/ethnicity representation) and demonstrate that the dataset reflects the intended use population. Include subgroup performance analysis showing the model performs equitably across demographic groups.

Common Pitfalls in FDA AI Submissions

Based on published FDA deficiency letters and industry experience, these are the most frequent training data issues that delay or derail AI device submissions.

  • 01.
    Data Leakage Between Splits. Patient-level data appearing in both training and test sets. This inflates performance metrics by 5-15% and is a red flag for FDA reviewers. Always split at the patient level, not the image level.
  • 02.
    Insufficient Demographic Diversity. Training on data from a single hospital or geographic region. FDA reviewers will question generalizability and may request additional validation studies.
  • 03.
    Undocumented Annotator Qualifications. Failing to document who labeled the data and their clinical credentials. "Trained annotators" is insufficient; FDA expects specifics on medical training, board certifications, and relevant experience.
  • 04.
    Missing Inter-Annotator Agreement. Not measuring or reporting how consistently annotators label the same data. Low agreement indicates noisy labels that undermine model reliability.
  • 05.
    Ignoring Edge Cases. Training only on clear-cut examples while the real world is full of ambiguous findings. FDA expects the model to handle borderline cases gracefully, which requires training data that includes them.

Pre-Submission Tips

The FDA's pre-submission (Q-Sub) process lets manufacturers meet with the agency before a formal submission to align on expectations. Here is how to make the most of it for AI/ML devices.

Prepare Early

File your pre-submission during the data collection phase, not after model development is complete. This gives you time to course-correct on data quality requirements, reference standard methodology, and clinical evaluation design before investing in full-scale annotation and model training.

Ask Specific Questions

Frame your Q-Sub around specific questions: "Is our proposed reference standard methodology adequate?" "Are our annotator qualifications sufficient?" "Is our dataset size and diversity adequate for the intended use?" Vague questions get vague answers.

Document Everything

Start documentation from day one: data provenance, annotation decisions, quality metrics, schema revisions. Retroactively reconstructing this documentation for a submission is time-consuming, error-prone, and often incomplete. Build documentation into your workflow, not as an afterthought.

Use Predicate Devices

Study FDA summaries of substantially equivalent devices already cleared in your space. Their training data methodology sets precedent for reviewer expectations. If similar devices used radiologist-labeled data with multi-reader consensus, plan your annotation accordingly.

How LabelCore Supports FDA Submissions

LabelCore.AI was built specifically for medical device companies navigating FDA requirements. Our annotation services include built-in regulatory documentation that integrates directly into your submission package.

  • Board-certified physician annotators with documented credentials, specialty training, and clinical experience relevant to your device's intended use.
  • Multi-reader consensus protocols with inter-annotator agreement metrics calculated and documented at every stage.
  • Complete audit trails with timestamped annotation logs, revision history, and adjudication records ready for FDA review.
  • FDA-format documentation packages including data source descriptions, annotator qualification summaries, and quality metric reports structured for 510(k) and De Novo submissions.

Frequently Asked Questions

Does the FDA require physician-labeled training data?+
The FDA does not explicitly mandate physician annotators, but their guidance documents emphasize that reference standards should reflect the clinical standard of care. In practice, FDA reviewers closely examine annotator qualifications during 510(k) and De Novo reviews. Submissions using board-certified physician labels consistently receive fewer deficiency questions about data quality compared to those using non-expert annotators.
What documentation does the FDA require for AI training data?+
The FDA expects thorough documentation including: data source descriptions, inclusion/exclusion criteria, demographic and clinical diversity metrics, annotation methodology and schema definitions, annotator qualifications and credentialing, inter-annotator agreement metrics, data splitting methodology (train/validation/test), and data preprocessing and augmentation descriptions. This documentation is typically submitted as part of the performance assessment section.
How does the FDA evaluate AI model performance?+
FDA evaluation focuses on clinical performance metrics relevant to the intended use (sensitivity, specificity, AUC, positive/negative predictive values) measured on a held-out test set that was not used during training or validation. The agency also examines subgroup performance across demographics, disease severity, and imaging equipment to ensure the model performs equitably across patient populations.
What is a Predetermined Change Control Plan (PCCP)?+
A PCCP allows manufacturers to describe anticipated modifications to an AI/ML device, including retraining on new data, and the methodology for implementing those changes without requiring a new submission for each update. The PCCP must specify what data quality standards will be maintained, how performance will be monitored, and what triggers a new regulatory submission. FDA finalized this guidance in 2024.

Need Expert Help?

Preparing training data for an FDA submission is high-stakes work. LabelCore.AI provides physician-grade annotation with built-in regulatory documentation designed for 510(k) and De Novo submissions.

Talk to Our Team