Written by Susan Miller*

Strategic English for Responding to FDA AI Requests: How to Write FDA Additional Information Responses for AI SaMD Without Overpromising

Facing an FDA Additional Information letter for your AI SaMD and unsure how to respond without overpromising? In this lesson, you’ll learn to craft regulator-ready replies that are precise, traceable, and calibrated—using a disciplined mapping structure, compliant phrasing templates, and bounded commitments that align with FDA expectations. You’ll find concise explanations, real-world examples and dialogues, and targeted exercises to lock in tone, structure, and timelines so your team speaks with one clear, defensible voice.

1) Understand the regulator’s intent and the tone FDA expects

When the FDA issues an Additional Information (AI) letter for an AI/ML Software as a Medical Device (SaMD), it is signaling three things: (1) specific gaps remain in your original submission, (2) the agency needs verifiable evidence to close those gaps, and (3) the agency expects a clear, accountable pathway to resolution. The intent is not adversarial. It is risk-targeted and patient-safety-oriented. The FDA wants to confirm that your device is safe and effective for the intended use, that your evidence is reproducible and traceable, and that your controls for ongoing model performance are robust. This intent shapes the tone: professional, precise, and grounded in objective data rather than persuasion.

Compared with EU MDR/Notified Body (NB) queries, FDA AI letters typically emphasize alignment with U.S. regulatory frameworks such as the QSR (or QMSR), guidance on Clinical Decision Support (CDS), Good Machine Learning Practice (GMLP) principles, and recognized consensus standards. While EU NB questions often track MDR Annex I General Safety and Performance Requirements (GSPRs) and focus on conformity assessments, FDA questions frequently aim to test the sufficiency of your methodological rigor and your change control strategy in the U.S. context. The tone for FDA is therefore slightly more evidentiary in structure and more prescriptive in how you document traceability between claims, datasets, metrics, and controls.

To meet this intent, your response tone should be:

  • Neutral, factual, and verifiable.
  • Directly cross-referenced to the FDA’s question numbers and sub-questions.
  • Transparent about what is known now and what will be produced, including timelines and dependencies.
  • Measured, avoiding speculative promises, but still accountable: state who will do what and by when, using cautious, testable commitments.

In short, the desired response characteristics are clarity, traceability, sufficiency of evidence, and calibrated commitments. Every sentence should help the reviewer locate evidence, understand rationale, and see a controlled path forward.

2) Use a disciplined response structure and a mapping technique

A point-by-point mapping converts a long AI letter into a series of small, high-confidence responses. This structure reduces ambiguity and helps reviewers verify resolution efficiently. The core structure includes: a cover note, a mapping matrix, the response body, and appendices.

  • Cover note: Briefly acknowledge receipt, state the submission number, and outline your organizational approach. Keep it short. Reaffirm your intention to address each point and to maintain traceability. Avoid advocacy language.

  • Mapping matrix: Create a table-like map (you may include as an appendix) that lists each FDA question identifier (e.g., Q1a, Q1b), the location of your response section, and the specific appendix where primary evidence sits. This map is the reviewer’s navigation aid. It also shows your internal discipline: each claim has a home, and each home has evidence.

  • Response body: For each question, restate or paraphrase the question to ensure alignment, then answer directly underneath. Use headings and subheadings that match the FDA’s numbering. Within each response, include: a concise claim or conclusion; the supporting rationale; key metrics; references to appendices; and, if needed, a clearly labeled “Next Steps and Commitments” subsection.

  • Appendices: Place detailed studies, protocols, raw results summaries, validation plans, human factors artifacts, cybersecurity threat models, and change control artifacts in appendices. Keep the body readable by referencing Appendix labels. Ensure each appendix has version control, dates, and authors.

The mapping technique aims to eliminate reviewer friction. If the FDA asks, “Provide demographic stratification for sensitivity and specificity across key subpopulations,” your mapping should point the reviewer to a specific section (e.g., 2.3.b) and then to a precise appendix (e.g., Appendix D: Subgroup Performance Tables, Rev B). The reviewer should not have to infer where data resides. The less they search, the more confidence they have that your processes are controlled and auditable.

While building the structure, maintain evidence hierarchy:

  • Body text: Claims and summaries only.
  • Appendices: Full protocols, analysis plans, tables, and figures.
  • Cross-references: Exact page or table numbers inside appendices.

Finally, adopt consistent terminology. Use the same model name, version, intended use statement, and metric definitions throughout. Inconsistent labels can trigger new questions and reduce perceived control.

3) Apply calibrated commitment language that avoids overpromising

Calibrated language signals accountability without creating regulatory debt. It distinguishes between what is already true and what you plan to do, and it frames future actions with conditions, traceable timelines, and risk-based rationales.

Key principles:

  • Separate facts from intentions. Facts are present-tense, evidenced statements. Intentions are future-tense and include dependencies and controls.
  • Use bounded commitments. Define scope, trigger conditions, timelines, and acceptance criteria. Avoid open-ended promises.
  • Reflect feasibility and regulatory alignment. If a method is subject to FDA feedback (e.g., a protocol), commit to submit, not to execute a specific outcome.

Helpful language patterns:

  • For current state: “The submitted model (Version X.Y) was validated on Dataset Z with pre-specified endpoints. Detailed subgroup metrics are provided in Appendix B, Tables B-3 to B-7.”
  • For near-term actions: “We will submit the finalized protocol for FDA feedback within 30 calendar days of this letter, and will not initiate data collection until alignment is confirmed.”
  • For contingent actions: “If inter-reader variability exceeds the predefined threshold (>5% absolute deviation) during the usability study, we will revise the training materials and conduct targeted retraining prior to resubmission.”
  • For limits and boundaries: “This commitment applies to the current intended use and the locked model Version X.Y. Any expansion in indications or algorithmic scope will be submitted via the appropriate regulatory pathway.”
  • For deferral without deflection: “We acknowledge the request for prospective clinical evidence. We propose a staged approach, beginning with retrospective external validation per Appendix G. Pending FDA concurrence, we will initiate a prospective study as outlined in the draft protocol.”

Calibrated language preserves credibility. Avoid terms like “guarantee,” “always,” or “will achieve superior performance,” unless supported by robust, repeatable evidence. Prefer “target,” “plan,” “intend,” and “subject to FDA feedback” when evidence is in development. Make sure every commitment is tracked internally, with owners and timelines, to ensure future submissions are consistent with what you promised.

4) Use compliant phrasing templates for common AI/ML SaMD topics

For AI/ML SaMD, some topics recur in FDA AI letters. Below are phrasing templates that balance clarity and caution. Tailor numbers, datasets, and references to your device.

  • Training data characterization and bias control:

    • “The training dataset comprises N samples acquired from M sites between [dates], with inclusion/exclusion criteria defined in Appendix A. Data collection adhered to pre-specified sourcing and labeling procedures (Appendix A-2). Demographic and clinical distributions are summarized in Table A-5. We assessed dataset representativeness against the target U.S. population using [reference standard].”
    • “We evaluated potential bias by stratified performance across age, sex, and race/ethnicity subgroups, and by acquisition device families. The predefined non-inferiority margin for subgroup performance is specified in the Statistical Analysis Plan (Appendix C). Where lower bounds of confidence intervals underperform the aggregate metric by >Δ, we have proposed mitigation steps in the Postmarket Surveillance Plan (Appendix J).”
  • Model updates and change control (locked vs. learning models):

    • “The currently submitted model is locked (Version X.Y). Any future modifications will follow the algorithm change protocol in Appendix H, with change categories classified per risk to safety and effectiveness. Minor updates (e.g., bug fixes, non-substantive pre-processing adjustments) will be managed under internal QMS with documented verification and validation. Substantive model changes affecting indications, inputs, or performance claims will be submitted through the appropriate FDA pathway.”
    • “For a potential future adaptive learning approach, we propose a Predetermined Change Control Plan (PCCP) consisting of: (1) SaMD Pre-Specifications defining anticipated changes; and (2) an Algorithm Change Protocol outlining data governance, retraining triggers, and acceptance criteria. We request FDA feedback on the PCCP scope prior to implementation.”
  • Performance claims and clinical relevance:

    • “Primary performance endpoints (sensitivity, specificity, AUC) were pre-specified in Appendix C. Confidence intervals are calculated using [method], with multiplicity controls described in Section 2.4. Claims are limited to the validated population and acquisition conditions listed in Appendix B. We do not generalize performance to untested modalities or indications.”
    • “Clinical relevance is supported by a decision-analytic rationale (Appendix F) linking performance thresholds to anticipated clinical impact. We do not assert superiority relative to standard-of-care without direct comparative evidence.”
  • Human factors and usability for AI-assisted workflows:

    • “We conducted formative studies to refine user interface elements that present AI outputs, uncertainty indicators, and recommended next steps. The summative usability test plan (Appendix E) includes representative users, critical tasks, and failure mode analysis. We will incorporate inter-reader agreement analysis to assess how AI outputs influence decision-making.”
    • “Risk controls address potential overreliance on AI: prominent labeling, confidence display, guardrails for ambiguous outputs, and training materials emphasizing clinician judgment. Updates to labeling will follow 21 CFR Part 801 requirements.”
  • Cybersecurity and data integrity:

    • “The software bill of materials (SBOM) and vulnerability management process are provided in Appendix K. We follow authenticated update mechanisms with code signing and maintain monitoring for known CVEs. Data at rest and in transit are protected via encryption consistent with recognized standards. Threat modeling (Appendix K-3) identifies attack vectors relevant to model integrity and inference pipelines.”
    • “Any security patches are evaluated for impact on performance and safety. Where patches alter runtime dependencies that could influence model behavior, we conduct regression testing per Appendix H and document outcomes.”
  • Postmarket monitoring and change management:

    • “We will monitor real-world performance via predefined Key Performance Indicators (KPIs) and stability metrics (Appendix J). Trigger thresholds for investigation include statistically significant drift in input distributions or degradation beyond Δ relative to the locked model’s baseline. Findings will inform corrective actions and, when applicable, regulatory notifications.”
    • “User complaint handling and field corrective actions will follow our QMS procedures. We will trend events by site and modality to detect systematic issues. Any safety signal will be escalated through our risk management process (ISO 14971 aligned).”

These templates help you present information in an FDA-ready format without asserting more than you can defend. Replace placeholders with concrete data and cite exact appendices.

5) De-escalate through clarity: state what is known, what is being generated, and realistic timelines

Clarity reduces regulatory friction. Begin each response with what is already established: validated results, controlled processes, and current labeling. Then separate what is underway (e.g., an ongoing external validation) from what is planned (e.g., a future prospective study), and from what is contingent (e.g., the scope of a PCCP pending FDA feedback).

Use explicit, bounded timelines. Instead of “soon,” say “within 30 calendar days of FDA concurrence.” Identify decision gates: protocol submission, FDA feedback, initiation of data collection, interim analysis, and final report. Connect each gate to a deliverable and an acceptance criterion. If timelines depend on external factors (e.g., site IRB approvals), state the dependency and how you mitigate it (e.g., multiple sites pre-qualified, parallel submissions).

Also clarify scope boundaries. If your performance claims apply only to certain devices, acquisition settings, or populations, state this explicitly and note that extrapolation will require new evidence. If you do not have subgroup power for rare populations, acknowledge the limitation and present your plan to monitor postmarket data. This transparency builds trust and prevents overpromising.

Finally, keep a consistent alignment statement: “We are committed to resolving the remaining questions within a documented, risk-based framework, coordinating with FDA as needed to ensure safety and effectiveness.” This is firm about accountability but does not claim results you have not yet produced.

6) Practice and quality-check with a scenario-driven checklist

To make your response reproducible and low-risk, run an internal quality check before submission. Use a checklist that evaluates completeness, traceability, and calibration.

  • Alignment and structure:

    • Do response headings exactly match FDA question numbers and wording? Have you paraphrased only where necessary to show understanding, without altering intent?
    • Does a mapping matrix point each question to a specific response section and appendix? Are appendix identifiers stable and versioned?
  • Evidence sufficiency:

    • Do all claims have primary evidence? Are tables and figures labeled, with methods described and confidence intervals provided where applicable?
    • Are dataset descriptions complete (sources, timeframes, inclusion/exclusion, curation, labeling processes)? Are subgroup analyses presented with rationale for thresholds?
  • Language calibration:

    • Are present-tense statements limited to verified facts? Are future actions framed with dates, dependencies, and acceptance criteria?
    • Have you avoided absolute promises and promotional claims? Are boundaries and limitations clearly stated?
  • Regulatory coherence:

    • Are recognized standards and guidances appropriately referenced? Does the change control plan reflect FDA expectations for PCCP where relevant?
    • Are labeling statements consistent with validated claims and intended use? Are human factors findings integrated into labeling and training?
  • Risk and postmarket controls:

    • Are drift detection, complaint handling, and signal escalation pathways defined? Are triggers quantitative and connected to corrective actions?
    • Is cybersecurity addressed across development, deployment, and update processes, with SBOM and vulnerability handling documented?
  • Consistency and traceability:

    • Are model names, versions, and dataset identifiers consistent across the document? Are acronyms defined once and used consistently?
    • Do cross-references point to exact appendix pages or table numbers? Are all hyperlinks or references functional and accurate?
  • Submission readiness:

    • Is there a concise cover note? Are all proprietary designations and confidentiality markings applied correctly?
    • Are responsible owners identified for each commitment, with internal tracking to ensure follow-through on future submissions?

Using this checklist ensures that your AI letter response is organized, verifiable, and responsibly scoped. It demonstrates to the FDA that your organization manages evidence and risk with discipline. Most importantly, it helps you maintain a professional, precise tone—one that advances the review while protecting you from overpromising.

By framing regulator intent, enforcing a point-by-point structure, using calibrated commitment language, employing compliant templates for AI/ML SaMD topics, and applying a rigorous quality checklist, you create a repeatable method for responding to FDA AI letters. This method reduces uncertainty, builds reviewer confidence, and moves your submission toward clearance without unnecessary risk.

  • Match FDA’s evidentiary tone: be neutral, precise, and traceable; align claims with verified data and U.S. frameworks (QMSR/QSR, GMLP, CDS guidance, standards).
  • Use a disciplined structure with a mapping matrix: restate each question, answer directly with claims, rationale, metrics, and exact appendix references; keep full evidence in appendices with version control.
  • Apply calibrated commitments: separate facts from plans, set bounded timelines and acceptance criteria, note dependencies (e.g., “subject to FDA feedback”), and define scope limits (model version, intended use).
  • De-escalate through clarity: state what is known vs. in progress vs. planned, specify decision gates and realistic timelines, and outline risk controls for updates, bias, cybersecurity, postmarket monitoring, and change management.

Example Sentences

  • We will submit the finalized external validation protocol within 30 calendar days of this letter, and will not initiate data collection until FDA feedback is received.
  • The submitted model (Version 2.3) was validated on Dataset Z using pre-specified endpoints; subgroup metrics are provided in Appendix B, Tables B-3 to B-7.
  • Any future modifications will follow the Algorithm Change Protocol in Appendix H, with substantive changes submitted through the appropriate FDA pathway.
  • We acknowledge the request for prospective clinical evidence and propose a staged approach beginning with retrospective external validation per Appendix G, pending FDA concurrence.
  • Trigger thresholds for investigation include statistically significant drift in input distributions or degradation beyond 3% absolute relative to baseline, as specified in Appendix J.

Example Dialogue

Alex: The FDA asked for demographic stratification and our change control plan. How should I phrase the response without overpromising?

Ben: Start neutral and precise: restate the question, then say, “Subgroup performance for age, sex, and race/ethnicity is in Section 2.3b and Appendix D, Tables D-1 to D-4.”

Alex: And for future work on the prospective study?

Ben: Use bounded commitments: “We will submit the finalized protocol within 30 days and will not begin enrollment until FDA concurrence; this commitment applies to the locked model Version 2.3.”

Alex: Should I mention adaptive updates?

Ben: Yes—propose a PCCP cautiously: “We request FDA feedback on the PCCP scope; anticipated changes and acceptance criteria are defined in Appendix H.”

Exercises

Multiple Choice

1. Which response best reflects the FDA-appropriate tone and structure when replying to an AI letter about subgroup performance?

  • “Our model clearly outperforms competitors in all demographics; please approve.”
  • “Subgroup performance for age, sex, and race/ethnicity is presented in Section 2.3b and Appendix D, Tables D-1 to D-4.”
  • “We promise to fix any issues as soon as possible.”
  • “See our website for detailed stats.”
Show Answer & Explanation

Correct Answer: “Subgroup performance for age, sex, and race/ethnicity is presented in Section 2.3b and Appendix D, Tables D-1 to D-4.”

Explanation: FDA expects neutral, verifiable, and traceable responses. Citing exact sections and appendix tables aligns with the mapping technique and evidentiary tone.

2. Which sentence uses calibrated commitment language that avoids overpromising?

  • “We guarantee superior performance across all hospitals.”
  • “We will submit the finalized protocol within 30 days and will not begin enrollment until FDA concurrence.”
  • “We will achieve AUC > 0.98 in every subgroup.”
  • “We will update the model continuously without further FDA interaction.”
Show Answer & Explanation

Correct Answer: “We will submit the finalized protocol within 30 days and will not begin enrollment until FDA concurrence.”

Explanation: This statement is bounded, includes a timeline and dependency on FDA feedback, and avoids absolute promises—matching the lesson’s guidance on calibrated commitments.

Fill in the Blanks

The response body should restate each FDA question, provide a concise claim, supporting rationale, key metrics, and cross-references to ___ where the primary evidence resides.

Show Answer & Explanation

Correct Answer: appendices

Explanation: Per the disciplined structure, the body contains claims and summaries, while appendices hold full protocols, analyses, and tables.

To de-escalate, the submission should clearly separate what is known, what is being generated, and realistic ___ tied to decision gates and deliverables.

Show Answer & Explanation

Correct Answer: timelines

Explanation: Stating explicit, bounded timelines linked to gates reduces ambiguity and aligns with FDA’s expectation for accountable planning.

Error Correction

Incorrect: We will always achieve superior performance and begin data collection immediately.

Show Correction & Explanation

Correct Sentence: We plan to submit the finalized protocol within 30 calendar days and will not initiate data collection until FDA feedback is received.

Explanation: Original uses overpromising (“always achieve”) and lacks dependencies. The correction applies calibrated commitment language and a clear dependency on FDA feedback.

Incorrect: See Appendix for details.

Show Correction & Explanation

Correct Sentence: See Appendix D (Tables D-1 to D-4) for subgroup sensitivity and specificity by age, sex, and race/ethnicity.

Explanation: Original lacks traceability. The corrected version provides precise appendix and table references, matching the mapping technique for verifiable navigation.