Tech

Why Healthcare AI Is Different: Building for HIPAA, CMS, and Patients — Not Just for Demos

After a decade building AI inside one of the largest U.S. health insurance ecosystems, Tirumala Ashish Kumar Manne unpacks what separates healthcare AI demos from production: HIPAA, CMS interoperability mandates, scale, security, and organizational readiness.

May 15, 2026
.
8 min read

Every few months, a new AI demo makes the rounds in healthcare circles. A model reads a clinical note and extracts diagnoses. A chatbot triages patient symptoms with impressive accuracy. A vision system flags anomalies in imaging scans. The excitement is genuine, and so is the underlying potential. But somewhere between the demo environment and a production system touching real patient data, something quietly breaks. Not the model itself. The architecture around it.

I have spent the better part of a decade building AI-enabled systems inside one of the largest health insurance ecosystems in the United States. What I want to offer here is not a critique of healthcare AI enthusiasm - I share it - but a practitioner's account of what actually changes when you move from a working prototype to a system that has to run at scale, in a regulated environment, for real patients. The differences are not incremental. They are structural.

The Regulatory Layer Is Not an Add-On

The first thing that catches prototype builders off guard is the regulatory environment. HIPAA is well known by name, but its operational implications are less widely understood. Protected health information cannot simply be passed to a third-party model API without a signed Business Associate Agreement in place. Audit trails are not optional logging - they are a compliance requirement. Every data access event touching a member record needs to be logged, attributed, and retrievable on demand. If your architecture does not build this in from the outset, you are not hardening a prototype. You are rebuilding it.

CMS interoperability mandates add another layer of complexity. The CMS Interoperability and Prior Authorization Final Rule, finalized in 2024, requires payers participating in Medicare Advantage, Medicaid, and CHIP to expose prior authorization data through standardized FHIR APIs. That means your AI system is not operating in isolation. It is expected to integrate with a regulatory API infrastructure that carries specific data format requirements, response time obligations, and its own audit trail. A prototype that does not account for FHIR R4 compliance from day one will spend months in retrofit before it can touch a production environment. I have watched that happen more than once.

The lesson is not that regulation makes innovation impossible. It is that in healthcare, regulation defines the shape of the system before you write a line of code. Architects who internalize this early tend to build faster in the long run, because they are not undoing decisions that were made without this context.

Scale Changes What Failure Means

In a demo, a model that is right ninety percent of the time is impressive. In a production system processing tens of millions of prior authorization decisions annually, that ten percent error rate is a clinical and operational problem with real consequences for real patients. A denial that should have been an approval means a patient waits for care. A processing delay means a provider is on the phone with a member who is already anxious. The stakes of inference accuracy in healthcare are not abstract, and no amount of impressive benchmark performance changes that.

This is why production healthcare AI systems require something that demo environments rarely have: a human-in-the-loop architecture that is thoughtfully designed from the start, not bolted on after the fact. The AI does not replace the decision. It informs it. The system has to be built to surface uncertainty, escalate edge cases, and maintain a clear record of which decisions were machine-assisted and which were human-reviewed. Designing that handoff cleanly is one of the harder architectural problems in this space, and it does not come up in a hackathon.

The other dimension of scale that prototype builders rarely contend with is the upstream dependency problem. Production healthcare platforms do not operate on clean data. They consume feeds from dozens of payer systems, provider directories, pharmacy benefit managers, and state Medicaid agencies, each with its own data formats, transmission schedules, and quality variability. Your AI system is only as reliable as the data pipeline feeding it. Building resilience into that pipeline, including deduplication logic, schema validation, and graceful degradation when a feed arrives late or malformed, is as important as model selection. It is also invisible in a demo, which is precisely why it catches teams off guard in production.

Security Cannot Be a Phase Two Problem

Healthcare data is among the most sensitive personal information that exists. It is also among the most targeted. The Change Healthcare breach in early 2024 disrupted claims processing across the United States for weeks, affecting hospitals, pharmacies, and patients in ways that made visible exactly how critical this infrastructure has become. The lesson for AI architects is direct: security posture is not a feature you add after the system works. It is a property of the architecture from the first design conversation.

In practice, this means adopting Zero Trust Architecture rather than relying on perimeter-based security assumptions. Every service call is authenticated. Every access is authorized based on identity and context, not network location. Encryption is applied at rest and in transit, consistently and without exception. Threat modeling happens at the design phase, not at the penetration testing phase. For AI systems specifically, it means thinking carefully about where model inference happens. A cloud-hosted third-party model endpoint represents a very different security posture than an on-premises or private cloud deployment when the inputs are protected health information.

These are not exotic requirements reserved for large institutions. They are baseline requirements for any system handling protected health information at scale. But they represent a significant architectural investment that a prototype simply does not carry, and teams that underestimate this tend to learn it the hard way.

What Actually Transfers from Prototype to Production

None of this is an argument against hackathons or rapid prototyping. The ideas that emerge from constrained, creative environments carry real value. The question is which parts of a prototype are worth carrying forward.

What transfers well is the problem framing, the user workflow insight, and the proof that a given AI capability is technically feasible for a given clinical or operational task. A hackathon team that demonstrates a working prior authorization suggestion model has validated something meaningful. That validation has real worth in the design process that follows.

What does not transfer is the data handling approach, the security model, the deployment architecture, and almost certainly the vendor integrations. These need to be rebuilt for the production environment, without exception. The teams that navigate this transition successfully treat the prototype as a hypothesis, not a foundation. They document what they learned about the problem, the data characteristics, and the model behavior, and they hand that documentation to an engineering team that builds production infrastructure from scratch with those insights as inputs. When prototype teams try to scale what they built at a hackathon, they typically spend more time on remediation than they would have spent building clean from the beginning.

A Note on Organizational Readiness

One dimension that rarely gets discussed outside of enterprise healthcare is organizational readiness. A technically sound AI system deployed into a clinical or operational workflow that the organization has not prepared for will underperform, regardless of how well the architecture is built. Clinicians and operations staff need to understand what the system does, what it does not do, and what it means when the system flags uncertainty. Building that understanding requires change management, training, and structured feedback loops. None of those appear in a model training pipeline.

The organizations that deploy healthcare AI successfully treat it as a cross-functional program, not a technology project. The architects and engineers are one part of that. Clinical informatics, compliance, legal, and operations are the other parts. Getting those teams in the room at the design stage is not a bureaucratic cost. It is what separates a system that gets adopted from one that runs perfectly in staging and quietly fails in production because nobody changed how people work around it.

The Opportunity Is Real

I want to close where I started: the excitement about AI in healthcare is warranted. The problems worth solving are genuinely important. Reducing prior authorization delays, surfacing clinical decision support at the right moment, making sense of fragmented patient records - AI has a meaningful role to play in all of them. The path from prototype to deployed system is harder than it looks from the outside, but it is a tractable engineering problem. The practitioners who understand that path clearly are the ones who will actually move the field forward.

If you are building in this space, I hope this is useful. And if you are building a healthcare AI prototype right now at a hackathon - keep going. Just know what comes next, and plan for it.

Related Blogs

No items found.