The 6–12% margin in your claims
Taxonomy, resubmission rules, and the agent loop that keeps it closed.
Architecture, human-in-the-loop design, and the SLAs we commit to. A ground-truth look at how we build agents that clinical leaders will actually trust.
A production pharmacy agent is not an LLM wrapper. It is a five-layer system — interface, orchestrator, tool library, knowledge base, integration fabric — governed by five guardrails and three SLAs (99.5% uptime, P95 response under four seconds, hallucination rate below half a percent). This piece is the architecture reference we use on every shipped agent, plus the six-week path from brief to production.
Most "pharmacy AI" demos collapse the moment they hit production. A pharmacist asks a clarifying question, the agent fabricates a drug interaction, leadership pulls the plug, and the vendor goes back to the drawing board for another six months. We've watched this cycle three times across the GCC in the last two years.
The problem isn't the model. Frontier models are more than capable. The problem is architecture — specifically, what a production agent has to do around the model to make it safe, auditable, and useful in a regulated clinical environment.
This piece walks through the architecture we deploy, the guardrails that keep it honest, and the service levels we commit to before we ask a pharmacist to rely on it.
An agent is not a chatbot. It is not a summarizer. It is not a "copilot" that suggests things you then copy-paste. Those things are useful, but they are not what we mean by an agent.
An agent, in our definition, has four properties:
If one of those four is missing, you do not have an agent. You have an assistant, and you should price it accordingly.
Here is the stack we run. Top to bottom:
The single biggest decision in pharmacy agent design is where the human sits in the loop. Get it wrong and you either (a) have an assistant that adds friction without removing work, or (b) you have a liability waiting to happen.
Our rule is simple: the human approves every irreversible action. Everything else — gathering data, drafting documents, filing evidence, updating internal dashboards — the agent does on its own.
Concretely, for each of the four practice areas:
The agent is fast. The human is accountable. That split is non-negotiable.
Five guardrails we build into every deployment:
When we deploy an agent into production, we sign an SLA with specific, measurable commitments. A representative one:
If we miss an SLA, the fee structure has teeth. This isn't vanity — it's how you force yourself to build the thing correctly the first time.
Most pharmacy operators expect an AI deployment to take a year. Ours are in production in six weeks. The phases:
By week 8, the agent is typically handling 60–80% of the drafting workload for its scoped function. The pharmacist's time goes to the judgment calls, which is where it should have been all along.
Not every workflow should be agentic. Simple automations don't need an LLM. Rule-based tasks don't need retrieval. If someone is trying to sell you an agent for something a well-configured PMS already handles, walk away.
The test we apply: does the task require reading unstructured text, making a judgment call across multiple data sources, or drafting a document? If yes, an agent earns its keep. If no, traditional software is faster and cheaper.
If you want to see the architecture running on your data, the diagnostic ends with a working prototype of one agent on one workflow. No slideware.
Taxonomy, resubmission rules, and the agent loop that keeps it closed.
Turning JCI/CBAHI from a fire drill into a steady-state system.
Data-backed leverage at the negotiation table.