
Abstractย
Medical AIย requiresย architectural principles that recognize medicine’s unique epistemological characteristics: informationย validatedย through survival outcomes, decisions affecting individual patients under irreducible particularity, and human threat-detection refined through evolutionary pressure. Weย establishย the liability rule as a boundary principle,ย demonstrateย why medical knowledge’s binary validation creates different requirements than domains with quantifiable quality metrics, and show how domain-specific contradiction detectionย maintainsย knowledge integrity. Recent stress-testing of frontier modelsย validatesย our predictions: systems trained on unconstrained data exhibit brittleness that information architecture prevents. Medical AI should augment physician practice, notย attemptย autonomous medical decisions.ย
Introductionย
Theย previously introducedย constrained competenceย framework proposed that reliable AI in high-stakes domainsย emergesย from information quality control rather than processing sophistication. This paperย operationalizesย that framework specifically for medical AIย systems supportingย diagnosisย and treatment of individual patients.ย
Medicine presents unique requirements because medical information quality isย validatedย through the ultimate binary outcome: patient survival. When medical knowledge is wrong, people die. This creates natural selection pressure on information thatย doesn’tย exist for most domains. Bad medical textbooks stop being used because people die. Good medical knowledge persists because it keeps people alive.ย
Microsoft researchers recently stress-tested six frontier models across medical benchmarks and found that high scores mask fundamental brittleness: modelsย maintainย accuracy when images are removed from visual diagnosis tasks, collapse under trivial perturbations like reordering answer choices, and generate fabricated medical reasoning. These failures stem from training on unconstrained information sourcesย the architectural choice our framework rejects.ย
The Binary Nature of Medical Knowledgeย
Medical knowledge differs fundamentally from other information domains because it isย validatedย through survival outcomes. Cardiac arrest protocols exist in their current form because alternative approaches resulted in worse survival rates. Each recommendation was debugged through millions of clinical encounters where wrong answers meant death.ย
This evolutionary pressure on medical knowledge creates specific AI requirements:ย
Source quality is non-negotiable: Medical AI must restrict sources to thoseย validatedย throughย culturally specific survival outcomes,ย peer-reviewed literature, evidence-based guidelines,ย andย systematically reviewed protocols. Unlike domains where AI can learn from diverse sources andย weightย them appropriately, medical AI must exclude information thatย hasn’tย beenย validatedย through clinical outcomes.ย
Uncertainty must be preserved: When medical evidence conflicts, systems cannot resolve ambiguity through probabilistic reasoning over unreliable sources. They must present uncertainty explicitly so physicians can exercise judgment.ย
Failure modes are catastrophic: Medicalย AI thatย generatesย plausibleย but incorrect informationย doesn’tย produce “lower quality” output.ย Itย creates conditions where people die.ย
The Velociraptor Principleย
Human physiciansย possessย threat-detection capabilities refined through millions of years of evolutionary pressure. When clinicians reportย that “something feels wrong” about a patient despite normal vital signs,ย they’reย accessing sensory systems that have been honedย through survival pressure. Organisms that missed threat signalsย didn’tย reproduce.ย
Humansย possessย approximately 1011ย sensory cells constantly sampling the environment. A physician’s discomfort with a patient presentation, a nurse’s intuition about deterioration, a surgeon’s tactile sense that tissue “feels wrong”:ย these are sophisticated pattern recognition systems debugged over millennia, not mystical abilities.ย
AI systems have never wrestled velociraptors for dinner. Theyย haven’tย experienced selection pressure that refined human threat detection. Until AI faces evolutionary consequences for errors, humans must remain between algorithmic recommendations and patient care. This recognitionย groundsย why certain medical judgments cannot be delegated.ย
The Liability Rule: Where AI Ends and Medicine Beginsย
We propose the liability rule as the boundary between permissible AI applications and those requiring human judgment:ย
Any AI output that couldย result in medical malpractice liability if incorrect must be reviewed and approvedย by a human physician.ย
This rule is both pragmatic (liability follows decision authority) and philosophical (medical decisions require judgment about irreducible particularity, such as this patient’s values, circumstances, andย embodied reality).ย
Applicationsย
Drug interaction alerts: “Drug A + Drug B has documented interaction per [citation]” = information retrieval. The physician decides whether to prescribe,ย given this patient’s situation. No liability transfer.ย
Differential diagnosis retrieval: “Symptoms {fever, cough, chest pain} appear in literature associated with pneumonia, pleurisy, pulmonary embolism. See: [citations]” = pattern matching over validated sources. The physician examines theย patient and makes a diagnosis. No liability transfer.ย
Autonomous diagnosis crosses the boundary: “This patient has pneumonia” = medical determination. If wrong and acted upon, it generates liability.ย
The pattern: systems retrieve, synthesize, and present information from validated sources. They do not make medical judgments about individual patients.ย
Medicine Is Not Healthcareย
This paper addressesย medical AI, not healthcare AI. The distinction matters:ย
Medicine: Diagnosing and treating individual patients. Binary outcomesย validatedย through survival. Irreducible particularity. Liability-generating decisions. Oftenย irreversible. Requires embodied judgment debugged through evolution.ย
Healthcare: Systems supporting medical practice but not directlyย determiningย individual patient outcomes. Resource allocation, scheduling, supply chains, population surveillance, qualityย metrics. Mistakes affect efficiency rather than mortality. Decisionsย generallyย reversible.ย
The liability rule applies strictly to medicine.ย A schedulingย AI that makes appointments inconvenient is annoying. Medical AI that makes diagnostic errors is lethal.ย
Domain-Specific Knowledge Integrityย
Medical knowledgeย evolves.ย Maintainingย knowledge base integrity requires detecting when new evidence contradicts existing content.ย
Why Domain Specialization Worksย
We propose domain-specific small language models (cardiology, pediatrics, oncology, emergency medicine) rather than general contradiction detection:ย
Constrained problem space: A cardiology model trained exclusivelyย onย cardiovascular literatureย doesn’tย attemptย to detect contradictions across allย medicine. Training data can be carefully curated from major cardiology journals, guidelines, and landmark trials.ย
Narrower task: These agents detect semantic contradiction, not medical truth. “Statement A and Statement B appear to make conflicting claims” โ judging which is correct. Human specialists make that determination.ย
Feasible training: Agents train on knownย contradictionย patterns. Beta-blocker recommendations for heart failure evolved from “contraindicated in acute HF” (1990s) to “mortality benefit in chronic HF” (2000s). Agents learn to recognize when new evidence challengesย establishedย positions.ย
Efficient specialist review: Flagged contradictions route to domain experts who resolve them in minutes. A cardiologist reviews cardiology flags,ย andย a pediatrician reviews pediatric flags.ย
Low-Threshold Flaggingย
Domain agents deliberately use low thresholds for flagging. If the pediatric agentย identifiesย aย possible conflictย between the 2019 AAP guidelines and theย 2023 Cochrane review on fever management, it flags “human attention needed” without judging which is correct.ย
Cost structure justifies this:ย
- False positive: Specialist reviews non-contradiction, spends 2-3 minutes, moves on. Cost: minimal.ย
- False negative: Realย contradictionย undetected, conflicting informationย remains,ย physicians receive inconsistent guidance. Cost: patient harm, eroded trust.ย
Resolution Workflowย
- Specialist receives flagged statements with contextย
- Confirms whetherย aย genuine contradiction existsย
- Determinesย which source is more current/authoritativeย
- Updates knowledge base metadata (deprecate outdated content, add qualifiers, flag “evolving evidence”)ย
- Resolution tracked and versionedย
This mirrors how medical knowledge evolves in practice. Guideline committees notice contradictions, review evidence, updateย recommendations. Constrained competence automates only the “noticing”. Specialistsย adjudicateย truth.ย
Leveraging Existing Hierarchiesย
Clinical guidelines alreadyย containย credibility scores:ย
- Grade A: Strong recommendation, high-quality evidence (multiple RCTs)ย
- Grade B: Moderate recommendation, moderate evidenceย
- Grade C: Weak recommendation, limited evidenceย
Theseย mapย directly toย constrainedย competence hierarchies. Medical AI preserves and surfaces these distinctions rather than creating themย de novo.ย
Empirical Validation: Why Unconstrained Systems Failย
Microsoft’s stress-testing revealed systematic brittleness in frontier models:ย
Modality shortcuts: Modelsย maintainedย 60-80% accuracy on visual diagnosis questions, even with images removed,ย relying on textual patterns and memorized associations rather than understanding.ย
Format brittleness: Reordering answer choices caused 4-6 percentage point drops. Models learned positional biases rather than medical content.ย
Distractor dependence: Replacing familiarย incorrect answers with irrelevant alternatives led to performance approachingย random guessing. Models rely onย eliminationย heuristics, not medical reasoning.ย
Visual-label shortcuts: Substituting images to align with distractor answers (text unchanged)ย resulted in drops of more than 30 percentage points. Modelsย learnedย shallow associationsย rather thanย robust integration.ย
Fabricated reasoning: Whenย prompted forย explanations, models generated plausible but incorrect justifications, hallucinated findings, or paired correct answers with invalid logic.ย
Why Constrained Competence Avoids These Failuresย
These failures stem fromย training onย unconstrained data containing misinformation alongside validated sources. Constrained competence avoids them architecturally:ย
No autonomous diagnosis: Systemย doesn’tย answer “What is the diagnosis?” It reformulates: “What does literature say about conditions with these symptoms?ย [citations]” Theย physician diagnoses.ย
Explicit boundaries: When visual information isย requiredย but unavailable: “Visual examination required for diagnosis.ย Literature on differential: [citations]” rather than guessing.ย
Source-grounded: All outputs referenceย validatedย sources. Cannot hallucinate treatments or findings absent fromย theย curated knowledge base.ย
Format independence: Retrieves information from sources rather than choosing among options. Reordering distractorsย isย irrelevant.ย
Uncertainty preservation: Surfacesย evidenceย quality distinctions (Grade A vs. C) rather than uniform confidence.ย
Appropriate refusal: When asked questions requiring medical judgment: “This requires clinical judgment about a specific patient.ย I can provide literature on [topic], but you make the medical decision.”ย
Implementation Considerationsย
Governanceย
Sourceย selection: Institutional committeesย establishย which sources enterย theย knowledge base (peer-reviewed journals, evidence-based guidelines,ย validatedย references)ย with explicitย exclusion of preprints (except emergencies), blogs, forums,ย andย non-peer-reviewed content.ย
Credibility assignment: Follow established frameworks (ACC/AHA grading, GRADE system, USPSTF ratings). Preserve rather than create hierarchies.ย
Update processes: Medical librarians with domain specialistsย monitorย major journals, updateย theย knowledge base on regular schedules.ย
Contradiction resolution: Domain agents flag conflicts, route toย appropriate specialistsย with accountability structures.ย
Integrationย
Clinical workflow position: Function as sophisticated reference tools consultable by physicians, not autonomous decision support requiring workflow adaptation.ย
Clear boundaries: Interface states: “This system retrieves peer-reviewed literature.ย You remain responsible for all medical decisions.” Reinforcesย theย liabilityย rule architecturally.ย
Scope Transparencyย
Domain coverage: Explicitly communicate what the system covers. “Includes: cardiology, pulmonology, emergency medicine. Excludes: sports medicine, occupational health.”ย
Query refusal: When asked questions requiring medical judgment: “This requires clinical judgment.ย I can provide literature on [topic], butย the treating physician must make the medical decision.” Refusal is a feature enforcing theย liabilityย rule.ย
Extending Beyond Medicineย
The architectural principles generalize to domains where decisions carry severe consequences and information is verifiable through rigorous outcomes:ย
Aviation safety: Maintenance proceduresย validatedย through catastrophic failure or successful operation. Domainย agentsย flagย contradictions;ย licensed mechanics decide.ย
Nuclear safety: Operational proceduresย validatedย through incidents. Licensed operators between AI recommendations and decisions.ย
Legal practice: Hierarchical credibility exists (Supreme Court > appellate > trial), but adversarial interpretation means multiple “correct” answers may coexist.ย The frameworkย must preserve conflicting authoritative sources. Licensed attorneys bear responsibility.ย
Unsuitable domains: creative endeavors (no authoritative hierarchies), exploratory research (benefits from unexpected connections), personal decisions (no binary validation or professional liability).ย
Domains suitable for constrained competence share: verifiable information hierarchies, professional liability structures, binary or near-binary validation through severe consequences,ย andย existing documentation practices.ย
Conclusionย
Medical AI’s path to reliability runs through information architecture, not processing power. Binary validation through survival outcomes creates requirements distinct from domains with quantifiable but non-fatal metrics. The liability ruleย establishesย that humans make medical decisions; AI retrieves survival-validated knowledge informing those decisions. Domain-specific contradiction detection with specialist reviewย maintainsย knowledgeย integrity. Empirical evidenceย validatesย these architectural choices: unconstrained systems failย robustnessย tests despite benchmark success.ย
Until AIย has wrestledย velociraptors for dinner, humans makeย the medicalย decisions. AI retrieves the knowledge from sourcesย validatedย through the ultimate test: patient survival.ย
Referencesย
Gu, Y., et al. (2025). The Illusion of Readiness: Stress Testing Large Frontier Models on Multimodal Medical Benchmarks.ย Microsoft Research, Health & Life Sciences.ย
Ferguson, J. (2025). Beyond prediction and explanation: Constrained competence as a third path for artificial intelligence in high-stakes domains. AI Journal.ย https://aijourn.com/beyond-prediction-and-explanation-constrained-competence-as-a-third-path-for-artificial-intelligence-in-high-stakes-domains/ย
Ferguson J. The Human Element: Ethical Guardrails for AI in Modern Medicine.ย The American Journal of Cosmetic Surgery. 2025;42(3):149-154.ย doi:10.1177/07488068251359686ย



