
In legacy industries like healthcare, the biggest opportunities live in “messy” problems where outcomes hinge on trust, empathy, and timing. In the age of Agentic AI, the temptation is to solve these with pureย compute. But automation aloneย isnโtย enough, and human-driven processesย don’tย scale.ย
Ifย youโreย building in this space, your job is to reconcile two competing truths:ย
- The importance of a human touch in earning trust and driving actionย
- The reliance of agentic systems on high-fidelity, “expert-in-the-loop” data for effectivenessย
At Headway, we faced this tension while helping primary care physicians (PCPs) refer patients to behavioral healthcare. Here is how we moved from manual workflows to a scalable, agentic-ready model.ย
Human-in-the-Loop as a Data Engineย
We often hear the Y Combinator mantra: โDo things that donโt scale.โ In the nascent era of Agentic AI, thisย roughly translatesย to: โGet scrappy to capture high leverage, high value data that doesnโt yet exist.โย
When we launched our first referral productโa simple “easy refer” buttonโit looked like a success on paper. But within a month, 50% of doctors churned.ย They didn’t want a button; they wanted a handoff.ย They missed the social workers they used to know by name.ย
We responded by virtually embedding Licensed Clinical Social Workers (LCSWs) into our referral flow. It was unscalable, but it was also a goldmine for agentic learning. By having LCSWs handle the “messy” middle:ย
- Capturing Intent: We recorded why a social worker chose one therapist over another for a specific patient.ย
- Decision Logic: Every manual interventionย becameย a labeled data point for how to navigate complex insurance or clinical barriers.ย
The Lesson:ย Don’tย just use humans to bridge a gap; use them to build a decision-log that will eventually serve as the “world model” for your AI agents.ย ย
Identifyingย High-Leverage Decision Bottlenecksย
As we moved toward conversion, we realized that trust isย asymmetric. One “high-leverage” moment can power the entire funnel. We discovered that when a doctor said, โJohn is going to reach out to you; heโs a therapist I trust,โ response rates doubled, even if theย subsequentย text was automated.ย
In an agentic system, this is whereย Contextualย Injection happens. An agentย shouldn’tย just “act”; it needs to know whose authority it isย operatingย under. Byย identifyingย these leverage points, you can design agents thatย don’tย just “execute tasks,” but “inherit trust.”ย
Implementing Agentic Guardrailsย
As we began layering in automation (outreach texts, follow-ups, scheduling, etc.) we hit the “Uncanny Valley.” One patient sharply asked: โYouโre not even a real person, are you?โย
This taught us the importance of Guardrails and Transparency. In healthcare, an agent hallucinating a tone or a clinical factย isn’tย just a bug;ย it’sย a liability. We implemented three layers of guardrails:ย
- Semantic Boundaries: Ensuring the agent never strays into diagnostic territory.ย
- Deterministic Fallbacks: If the AI detects high emotional distress or a “logic loop,” it immediatelyย pingsย a human supervisor.ย
- Transparency by Design: We stopped trying to “mask” the AI. We told patients: โTechnology helps us move faster, but John (a real human) is monitoring this to ensure you get care.โย
Surprisingly, patientsย didn’tย mind us incorporating AI into our experiences; what they minded was any attempt to hide it.ย When the guardrails were visible, trust actually increased.ย ย
The Flywheel of Decision Dataย
Today, Headwayโs referral productย operatesย nationwide. But the real “moat”ย isn’tย just the network;ย it’sย the rich decision-making data captured from years of human-in-the-loop processes.ย ย
Building in vertical AI (especially in high-sensitivity fields) requires a shift in mindset:ย
- Phase 1: Human-in-the-loop for trust and data capture.ย
- Phase 2: Agent-in-the-loop for scale and efficiency.ย
- Phase 3: Human-on-the-loop for oversight and edge-case handling.ย
If you are building healthcare AI products,ย don’tย rush to remove the human. Instead, celebrate the outsized leverage of human processes to generate rich training data early on, and deploy humanย expertiseย to shape the nature of the decision making that will eventually be the domain of the agents you build.ย
ย
ย



