Handvantage

Healthcare and agentic AI: HIPAA, FDA SaMD, and the supervision question.

Three regulatory frames a healthcare buyer has to satisfy simultaneously, and the line between platforms that fit administrative use and platforms that require FDA clearance.

Feature image for "Healthcare and agentic AI: HIPAA, FDA SaMD, and the supervision question."

Healthcare is the sector where agentic AI procurement decisions take the longest, and the reason is structural: a healthcare deployment has to satisfy three overlapping regulatory frames simultaneously, none of which were designed with agentic AI in mind. The frames overlap differently for different deployment shapes — an AI assistant that drafts internal emails for hospital administrators sits in a very different regulatory posture from an AI tool that summarises a patient encounter for a clinician. This dossier walks through the three frames, the line that separates productivity tooling from regulated medical devices, and what a healthcare buyer should ask any AI vendor before procurement begins.

Before the regulatory walk-through: a critical scope clarification. Vantage Workspace is a productivity platform with strong audit posture. It is not a clinical decision support system. It is not a medical device under any FDA definition. It does not produce clinical recommendations, does not interpret medical images, does not generate diagnostic conclusions. The platform fits administrative and clinically-adjacent uses (internal communications, document drafting, file management, agent-assisted operational work) where the regulatory exposure is real but bounded. Where a healthcare organisation needs an AI tool that participates in clinical decision-making, that is a different product category — typically requiring FDA 510(k) clearance, ONC certification, or both — and we will tell you so in the first conversation. This dossier explains the line and how to think about it in procurement.

The first frame is HIPAA. The Health Insurance Portability and Accountability Act, with its Privacy Rule (45 CFR Part 164 Subpart E) and Security Rule (Subpart C), is the operative federal framework whenever Protected Health Information is involved. PHI is broadly defined: any individually identifiable health information held or transmitted by a covered entity or business associate, in any form, including electronic. The moment an agentic AI tool reads, drafts, summarises, or transmits PHI, the platform vendor becomes a Business Associate under 45 CFR 160.103, requiring a Business Associate Agreement, the Security Rule's administrative/physical/technical safeguards, breach notification under the HITECH amendments, and the minimum-necessary standard under 164.502(b).

The minimum-necessary standard is the part agentic AI buyers most often underestimate. Under 164.502(b), a covered entity must make reasonable efforts to use, disclose, and request only the minimum amount of PHI necessary to accomplish the intended purpose. Applied to an AI agent, this means: the agent should not have undifferentiated access to a patient's full record when the task at hand only requires a specific section. A chatbot that ingests an entire EHR record to answer a scheduling question fails the minimum-necessary standard, even if every other HIPAA control is in place. The structural answer is per-task scope enforcement at the platform level — which is what Vantage Workspace's per-agent permission model and policy engine produce, but which is missing from most agentic AI products that ship as “give the agent your data and let it figure out what it needs.”

The second HIPAA dimension that catches healthcare deployments: the breach notification calculus when an AI's output is itself the breach. If an agent's reply contains information from another patient's record (a memory leak, an embedding inversion, a retrieval scope error), the disclosure is reportable under 164.404. The notification clock starts at discovery. The HHS Office for Civil Rights has signalled in 2025 enforcement guidance that AI-driven incidents are a focus area, with several settlements involving AI vendors entering Resolution Agreements with corrective action plans. The platform's posture has to assume incidents will happen and produce the contemporaneous evidence record that makes notification fast, accurate, and bounded — not retrospective forensics that take weeks to scope.

The second frame is FDA Software as a Medical Device. The FDA's authority over software as a medical device derives from the Federal Food, Drug, and Cosmetic Act and is operationalised through the Center for Devices and Radiological Health (CDRH). The relevant guidance for AI/ML-enabled software has matured significantly since the FDA's 2019 discussion paper, with the January 2025 final guidance on Predetermined Change Control Plans (PCCPs) for AI/ML SaMD providing the current procedural framework for how such products are cleared and how they can change post-market without requiring re-clearance.

The line that matters for healthcare AI procurement is whether the software is a medical device under section 201(h) of the FD&C Act. The 21st Century Cures Act narrowed the definition for clinical decision support software at section 3060, exempting CDS that meets four specific criteria (administrative support functions; software that displays, analyses, or prints information; software that supports or provides recommendations to a healthcare professional about diagnosis, treatment, or disease prevention; and software where the healthcare professional is intended to be able to independently review the basis for the recommendation). Software that fails any of the four criteria — particularly the “independent review” clause for recommendation-generating tools — falls back into the device definition and requires FDA pathway selection.

Applied to agentic AI, the test is concrete: an agent that drafts a clinical note for a physician to review and edit before signing is administrative; an agent that generates a diagnostic recommendation that the physician relies on without independent verification is potentially device-regulated. The line is sharper than vendor marketing suggests. The 2025 guidance specifically addresses how iterative AI changes (model retraining, prompt template updates, tool catalogue changes) interact with the device pathway: a PCCP allows pre-authorised changes within a declared envelope; changes outside the envelope require new submission. For an agentic platform that updates frequently, the PCCP framework is the operational reality, not the exception.

What this means for procurement: a healthcare organisation evaluating Vantage Workspace for administrative and clinically-adjacent uses (internal communications, operational document drafting, scheduling coordination, vendor management) faces a HIPAA evaluation but not an FDA evaluation. A healthcare organisation evaluating an agentic AI product for clinical decision support faces both, and should be cautious of vendors who claim FDA exemption without explaining which of the four Cures Act criteria the product satisfies. We do not claim FDA exemption because we do not deploy in clinical decision support; the question does not arise.

The third frame is the supervision and accountability layer that surrounds AI use in healthcare even when neither HIPAA nor FDA is the gating constraint. State medical boards have begun publishing guidance on AI-assisted clinical practice, with California's Medical Board (under AB 3030, effective January 2025) requiring physicians to review and confirm AI-generated content before it is communicated to patients, and similar requirements emerging in Texas (HB 4660), New York (the proposed AI in Healthcare Disclosure Act), and the Federation of State Medical Boards' 2024 model guidance. The accountability remains with the licensed clinician; the AI tool is treated as a medical assistant whose work the clinician supervises and signs.

The Joint Commission's standards for hospital accreditation address AI under the Information Management chapter and the Patient Safety Systems chapter, with the 2025 revision adding specific expectations around AI governance, including documented policies for AI tool selection, validation processes for AI outputs in clinical contexts, and incident reporting for AI-related adverse events. Joint Commission surveyors increasingly ask to see the AI inventory and the supervision framework as part of standard surveys, not as a special focus area.

Liability allocation is the question that ultimately determines deployment shape. Under current malpractice doctrine, the supervising physician retains responsibility for clinical decisions even when those decisions are informed or generated by AI. A vendor that fails to deliver the audit trail necessary to demonstrate physician supervision is exposing the physician to a defence-burdening evidence gap. A platform that produces unified, signed, contemporaneous records of AI output and physician review is the platform that supports defensible practice. This is true regardless of FDA status; it is a malpractice insurance question and a state medical board question, and it does not go away because a vendor has FDA clearance.

The platform's posture for healthcare deployments. Vantage Workspace fits the administrative and clinically-adjacent layer of a healthcare organisation's workload — the work that touches PHI but does not generate clinical recommendations. The platform produces the HIPAA-required evidence record (event-level audit log, signed timestamps, breach-notification-ready exports), supports per-agent scope enforcement consistent with the minimum-necessary standard, ships with Keycloak preconfigured (federating to the customer’s existing identity provider — typically Okta, Microsoft Entra ID, or Epic-integrated identity in EHR-heavy environments — when one is in place), and feeds the customer’s existing SIEM where their security operations are already running.

What the platform does NOT do for healthcare: it does not generate clinical recommendations, interpret diagnostic data, or produce content that a clinician would rely on without independent review. It does not have FDA clearance because it does not require it. It does not replace the EHR — the platform integrates with EHR systems where the customer’s deployment requires it, but the EHR remains the system of record for clinical data. It does not handle the physician supervision workflow on its own — that workflow is configured by the customer’s medical informatics team using the platform’s policy framework.

What the customer still owns. The HIPAA management of the deployment is the customer’s; we sign a BAA that allocates platform-level safeguards. The minimum-necessary policies (which agent has access to which scope of PHI, for which task) are configured by the customer through the platform’s policy YAML — we provide the framework; the customer’s privacy office writes the rules. The breach notification process is the customer’s; we provide the evidence the customer’s privacy office needs to make the determination quickly. The clinical supervision framework — who supervises which AI use, what the review consists of, how the review is documented — is the customer’s medical leadership’s, not a configurable preference.

Three deployment patterns we see in healthcare procurement, ordered by regulatory complexity. The first is administrative-only: the platform is used for internal hospital operations (HR communications, vendor management, project documentation, scheduling coordination) where PHI is incidentally present but not the focus of the work. HIPAA applies; FDA does not; supervision is general organisational accountability. This is the most common deployment shape and the one with the shortest procurement cycle.

The second pattern is clinically-adjacent: the platform supports work that touches clinical operations without participating in clinical decisions. Examples include drafting internal communications about clinical policy, summarising operational metrics for medical leadership, preparing documentation for accreditation surveys, and coordinating multi-disciplinary team work across systems. HIPAA applies in full; FDA generally does not, provided no clinical recommendations are generated; supervision is the existing organisational hierarchy. Most healthcare deployments of agentic productivity platforms land here.

The third pattern is clinical-decision-support, where the AI participates in care decisions. This is where Vantage Workspace stops being the right product. A healthcare organisation evaluating clinical-decision-support AI should be talking to vendors who specialise in that category, can demonstrate FDA pathway compliance (510(k), De Novo, or PMA depending on risk class), and can support the clinical validation studies and post-market surveillance the FDA expects. We will tell a healthcare buyer this directly if they describe a clinical-decision-support use case in the first conversation; the alternative is to take a deal we cannot serve well, which is bad for the customer and worse for the healthcare ecosystem.

Four questions a healthcare buyer should ask any agentic AI vendor before procurement. First: “Will you sign a Business Associate Agreement, and which of the Security Rule’s administrative, physical, and technical safeguards do you implement at the platform level?” Vendors who hesitate at the BAA, or who provide a generic answer to the safeguards question, are not ready for healthcare deployment.

Second: “How does the platform enforce the minimum-necessary standard at the per-agent, per-task level? Show me a worked example of an agent’s permission scope being narrower than the user who delegated to it.” The right answer involves a policy mechanism that is enforced at runtime, not a documentation claim about role-based access controls. The wrong answer is “the agent inherits the user’s permissions” — that fails the minimum-necessary test.

Third: “Is your platform regulated as a medical device under FDA SaMD criteria? If you claim exemption under the Cures Act, which of the four section 3060 criteria does the platform satisfy, and how would you defend that classification to FDA?” Vendors who answer this clearly understand the regulatory landscape; vendors who answer evasively either don’t understand it or are hoping the question doesn’t come up.

Fourth: “What does the audit record look like for a single physician interaction across a 24-hour shift, and can you produce it during an OCR examination or a Joint Commission survey within the response window the regulator allows?” This question separates platforms that produce the unified interaction record from platforms that produce three disconnected logs the customer assembles after the fact.

The 2026 healthcare AI procurement environment is being shaped by three converging pressures: HHS OCR’s expanded AI focus in HIPAA enforcement, the FDA’s increasingly mature SaMD pathway making the device line clearer (and therefore harder for grey-zone products to live in), and state medical board guidance creating supervision expectations that operate independently of federal frameworks. The healthcare organisations that will deploy agentic AI successfully in this environment are the ones that match the right platform to the right use case, with eyes open about which regulatory frames apply and where the customer’s management system retains responsibility. The organisations that will struggle are the ones whose vendors are vague about the line between administrative and clinical, whose audit trails are reconstructed rather than contemporaneous, and whose AI inventory is missing the supervisor-of-record assignments that state regulators are starting to ask for.

Vantage Workspace fits a specific portion of the healthcare AI workload. The portion it fits is large — most administrative and clinically-adjacent productivity work in a hospital, payer, or life-sciences organisation falls in this category — and the regulatory burden is real but bounded. The portion it does not fit (clinical decision support) is a different product category. A healthcare buyer who knows which use cases sit where in this map can make a procurement decision faster, with less risk of either deploying the wrong tool for the job or accepting unnecessary regulatory exposure. The next conversation, if it’s warranted, is the one where we walk through your specific use cases and tell you which ones we serve well, which ones we don’t, and why.



CONTINUE THE CONVERSATION

If something here is what you're working on, talk to us.

Articles like this one come out of conversations with practitioners, security leaders, and engineering teams in regulated industries. If the writing reflects your situation, the next conversation is probably worth having.

Continue the conversation →

Or write to hello@handvantage.com directly.