TA1/TA2 Solution: Hard-to-Vary AI
Our ADVOCATE solution implements Popperian epistemology through a three-component architecture. Each component is named after a philosopher or mythological figure whose work embodies its function.
Generates hypotheses via conjecture-refutation cycles. Creates multiple competing explanations and scores them for 'hard-to-vary' quality.
Ensures epistemological metadata (HTV scores, evidence grades, falsification criteria) is structurally enforced between agents.
Evaluates safety boundaries and makes APPROVE/ROUTE/HARD_STOP decisions. Enforces falsifiability requirements.
Key Insight: Most healthcare AI optimizes for prediction accuracy. We optimize for explanation quality. Predictions can be right for wrong reasons. Explanations that are hard to vary remain correctable and auditable even when wrong.
TL;DR
Most healthcare AI systems optimize for prediction accuracy. We optimize for explanation quality.
The difference: predictions can be right for wrong reasons (and silently fail when conditions change). Explanations that are hard to vary where every component is load-bearing remain correctable and auditable even when wrong.
What We Implement
| Mechanism | Purpose |
|---|---|
| ArgMed Debate | Generate multiple hypotheses, attack each adversarially, keep only survivors |
| HTV Scoring | Quantify how "hard to vary" each explanation is (0.0-1.0) |
| IDK Protocol | 12 specific uncertainty triggers with structured responses |
| Falsification Criteria | Every claim specifies what would prove it wrong |
| Safety Routing | High-risk decisions require human clinician approval |
| Clinician Feedback Loop | Overrides actively change future reasoning for that patient |
| Composable Domains | Medication, nutrition, exercise, sleep, mental health: same principles |
| Rules as Data | Interaction rules are explicit, versioned, auditable data |
| Accuracy Ascertainment | We measure our own predictions against outcomes |
The result: an AI that tells you why, admits when it doesn't know, specifies how to prove it wrong, measures whether it was right, and learns from clinician corrections.
Philosophy Foundations
The fundamental failure of current health AI is epistemological, not computational. Most health AI makes predictions based on patterns. But pattern-matching is not understanding.
The Inductive Fallacy
| Failure Mode | Problem | Consequence |
|---|---|---|
| Correlation does not equal Causation | Pattern matching finds correlations, not causes | Interventions based on spurious correlations fail or cause harm |
| The Black Swan Problem | Rare cases don't match common patterns. Medicine is full of rare cases | Induction fails precisely on the edge cases that matter most |
| Easy-to-Vary Explanations | Probabilistic outputs are 'mushy': you can change details without breaking the theory | 'You might have A, B, or C' explains nothing and helps no one |
These are practical manifestations of the deeper Hume/Popper critique: induction (inferring general rules from particular observations) cannot justify knowledge. No amount of data can prove a universal claim, but a single counterexample can refute one.
The Seasons Example
In "The Beginning of Infinity" (Ch. 1), Deutsch illustrates good vs. bad explanations using the ancient Greek explanation for seasons:
Persephone, goddess of spring, was kidnapped by Hades. Her mother Demeter's grief causes winter. When Persephone returns, spring comes.
Could substitute any gods or emotions. Nothing is load-bearing.
Earth's axis is tilted 23.5 degrees relative to its orbital plane around the sun. This causes different hemispheres to receive more direct sunlight at different times of year.
Change the tilt angle and predictions break. Every detail constrains.
This distinction is the foundation of our entire architecture.
Our Epistemological Foundation
We build on two complementary philosophical frameworks:
Karl Popper
Conjecture and Refutation
Science advances not by confirming theories but by attempting to refute them. The demarcation between science and pseudoscience is falsifiability.
David Deutsch
Hard-to-Vary Explanations
Good explanations begin with bad explanations. You get there by criticism, by conjecturing variants, and choosing the one that survives.
Philosophy Made Concrete
| Philosophical Principle | Component | Implementation |
|---|---|---|
| Conjecture-Refutation | Deutsch ArgMed Debate | Multi-agent Generator -> Verifier -> Reasoner pipeline |
| Hard-to-Vary Criterion | HTV Scoring | 4-dimensional algorithm (interdependence, specificity, non-adhocness, falsifiability) |
| Boundary Enforcement | Popper Safety Rules | Deterministic policy engine enforcing safety boundaries |
| Fallibilism | IDK Protocol | Structured honest uncertainty admission |
| Falsifiability | FalsificationCriteria | Every claim includes explicit refutation conditions |
| Error Correction | Clinician Feedback Loop | Override tracking with confidence decay |
System Architecture
Core Components
Three interconnected mechanisms form the heart of our epistemological architecture:
ArgMed Debate: Conjecture-Refutation in Action
The ArgMed (Argumentative Medicine) debate is our core reasoning mechanism. It directly implements Popperian conjecture-refutation through a three-agent architecture:
Generator
Conjecturer
Produces multiple hypotheses spanning different mechanisms
Verifier
Critic
Attacks each hypothesis adversarially, scores HTV
Reasoner
Synthesizer
Selects survivors based on HTV threshold
Mechanism Diversity Requirement
✓ Good Differential (chest pain)
- Acute coronary syndrome (cardiac)
- Pulmonary embolism (pulmonary)
- Musculoskeletal pain (muscle/bone)
✕ Bad Differential (rejected)
- ST-elevation heart attack (cardiac)
- Non-ST-elevation heart attack (cardiac)
- Unstable angina (cardiac)
The second example is easy to vary within the cardiac category, all three share the same mechanism. This violates Deutschian epistemology.
HTV Scoring: Operationalizing "Hard to Vary"
The HTV (Hard-to-Vary) score quantifies how hard it is to vary an explanation while preserving its predictions. We score explanations on four dimensions:
| Dimension | Question | High Score | Low Score |
|---|---|---|---|
| Interdependence | How tightly coupled are the components? | Every piece connects to the conclusion | Components could be swapped |
| Specificity | How precise are the predictions? | Specific, measurable outcomes | Vague, unfalsifiable predictions |
| Non-adhocness | Are all elements load-bearing? | Removing any element changes predictions | Contains free parameters |
| Falsifiability | What would refute this claim? | Clear conditions that prove it wrong | Immune to counterevidence |
HTV Thresholds (v0.1)
Worked Example: Fatigue
"You feel tired because of stress."
"Your fatigue is caused by iron deficiency anemia. Ferritin 8 ng/mL indicates depleted stores. Hemoglobin 10.2 g/dL confirms anemia."
IDK Protocol: Fallibilism as a Feature
Acknowledging the limits of knowledge is a virtue, not a failure.
Expressing the Deutschian position on fallibilism
The IDK (I Don't Know) Protocol formalizes how our system handles situations where it cannot make a confident recommendation. When we trigger IDK, we're not claiming the problem is unsolvable. Deutsch's optimism states that all problems are soluble given the right knowledge.
7 Core Triggers
| Trigger | Condition | Default Action |
|---|---|---|
| IDK_HTV_LOW | Composite below 0.4 | Route to clinician |
| IDK_NO_SURVIVORS | All hypotheses rejected | Route to clinician |
| IDK_MISSING_SIGNAL | Critical data absent | Request more info |
| IDK_CONFLICT | Unresolved contradictions | Route to clinician |
| IDK_EVIDENCE_WEAK | Only expert opinion available | Route to clinician |
| IDK_STALE | Snapshot too old | Request refresh |
| IDK_OUT_OF_SCOPE | Query outside domain | Deflect appropriately |
The Discriminator: Breaking Ties Between Theories
When multiple theories survive with equal HTV scores, we don't guess. We identify the discriminator: the single test that would kill one theory but not the other.
| Competing Theories | Discriminator | Logic |
|---|---|---|
| Iron vs. B12 Deficiency | MCV | Iron: low MCV; B12: high MCV |
| Heart Failure vs. Venous Insufficiency | BNP Level | HF: elevated; Venous: normal |
| Hypothyroidism vs. Depression | TSH Level | Hypothyroid: elevated; Depression: normal |
End-to-End Example
What We're NOT Claiming
Intellectual honesty requires acknowledging limitations:
01 We're not claiming AI can "create explanatory knowledge"
In Deutsch's philosophical sense, explanatory knowledge emerges through genuine understanding: the ability to creatively vary conjectures and recognize when variations break the explanation. Large language models operate through sophisticated pattern matching on training data.
What we implement is structured reasoning that approximates conjecture-refutation. The Generator doesn't truly "understand" why a hypothesis explains the data; it produces outputs that structurally resemble good explanations.
The epistemological structure is real. The underlying cognition is not Deutschian knowledge creation.
02 Human oversight remains essential
The Popper routing mechanism exists precisely because we don't trust the AI to handle all cases. This is not a temporary limitation to be engineered away; it's a principled design choice.
High-risk decisions require human judgment because:
- Large language models can produce confident, well-structured, wrong outputs
- Medical decisions involve values and trade-offs beyond optimization
- Accountability requires a human decision-maker
Routing to clinicians is a feature, not a bug.
03 HTV doesn't fix model hallucination
A subtle point: a large language model can hallucinate a high-HTV explanation. It can fabricate specific lab values, invent plausible mechanisms, and generate falsifiable predictions, all of which are false.
HTV measures structural quality of explanation, not correspondence to reality.
Red-Team Example: Fabricated Specificity
"Patient has hypokalemia-induced arrhythmia risk due to K+ of 2.9 mEq/L from recent diarrheal illness, exacerbated by concurrent furosemide 80mg daily."
The problem: The K+ value was fabricated. Patient's actual K+ is 4.1 mEq/L.
This is why HTV operates alongside provenance verification, snapshot grounding, and clinician oversight, not as a standalone safety measure.
04 This is a methodological commitment
We're making a bet: that AI systems structured around epistemological principles will be more reliable, more auditable, and more correctable than systems optimized purely for prediction accuracy. This is not a claim about machine consciousness, understanding, or intelligence. It's a claim about architecture.
The Value Proposition
The value is not that our AI "thinks like Deutsch". It doesn't. The value is that by structuring outputs to include HTV scores, falsification criteria, evidence grading, and honest uncertainty, we create systems that are:
When wrong, we know why and how to fix it
Every decision has a traceable reasoning chain
Low confidence triggers routing, not overconfident action
Error patterns can be identified and addressed
Why This Matters
Your doctor stays in control
Medication changes are always reviewed and approved by your clinician
Explanations, not just predictions
You understand why a recommendation is made
Honest uncertainty
The system tells you what it doesn't know
Your data matters
Decisions are grounded in your specific situation, not generic advice
Medication proposals require your approval
AI proposes start/stop/titrate/hold; you decide
Audit trails with epistemological metadata
Every decision is reviewable with full reasoning chain
System admits uncertainty
No overconfident black boxes. Low confidence triggers routing to you
Your overrides matter
Rejections and modifications actively change future recommendations for that patient
The ability to create new explanations is the defining attribute of people.
David Deutsch, The Beginning of Infinity (Ch. 7)
In Chapter 7 ('Artificial Creativity'), Deutsch argues that genuine AI must involve explanation and creativity, not just prediction. This implies that AI systems handling high-stakes domains need mechanisms for generating and evaluating explanations, not just pattern matching.
Our Approach Prioritizes
Error Correction over Error Prevention
We assume we'll be wrong and build in correction mechanisms
Fallibilism over Certainty
We never claim final answers
Explanation over Prediction
We require every claim to be justifiable
Open Questions
Intellectual honesty requires acknowledging not just what we don't claim, but what we don't yet know. These are active research questions we're working through.
Glossary
Key terms used throughout this document. Click any highlighted term in the text to see its definition.
A score (0.0-1.0) measuring how much each part of an explanation is load-bearing.
What observations would prove a claim wrong.
Popper's method: propose ideas, then try to disprove them.
Structured admission of uncertainty with 12 specific trigger types.
How methodologically rigorous the supporting evidence is.
Information about the quality and basis of a claim.
A test designed to falsify one theory while leaving another intact.
A safety mechanism that automatically stops the system when errors become too frequent.
References
- Popper, K. (1963). Conjectures and Refutations: The Growth of Scientific Knowledge. Routledge.
- Deutsch, D. (2011). The Beginning of Infinity: Explanations That Transform the World. Viking.
- Taleb, N.N. (2007). The Black Swan: The Impact of the Highly Improbable. Random House.
- Hunter, A. & Williams, M. (2012). Aggregating evidence about the positive and negative effects of treatments. Artificial Intelligence in Medicine, 56(3), 173-190.
- GRADE Working Group (2004). Grading quality of evidence and strength of recommendations. BMJ, 328(7454), 1490.
- Regain Health (2025). Hard to Vary Specification v0.9.0. Internal documentation.
Prior Work & Influences
We have built upon the following work:
| Component | Based On | Our Extension |
|---|---|---|
| ArgMed Debate | ArgMed-Agents (Hong et al., 2024) | Added HTV scoring + Popperian falsification |
| Multi-Agent Debate | Du et al., 2023 | Applied to clinical domain with safety routing |
| Evidence Hierarchy | GRADE / Canadian Task Force | Reinterpreted through Deutschian "hard to vary" lens |
| HTV Operationalization | Parascandolo et al., ICLR 2021 | Extended to 4-dimension clinical scoring |
| Argumentation Schemes | Douglas Walton | Applied to clinical reasoning |
| Uncertainty Abstention | Leibig et al., 2019 | Formalized as IDK Protocol with trigger taxonomy |
| Hypothesis Diversity | Kammer et al., 2021 | Enforced via mechanism-diversity requirement |
What We Believe Is Novel
- • Unified architecture combining all above under Deutschian/Popperian philosophy
- • The Non-Trust Principle (epistemological metadata increases but never decreases conservatism)
- • Safety rules implementing demarcation as executable policy
- • Clinician feedback loop with confidence decay
- • Systematic application of Deutsch's philosophy to healthcare AI architecture
Ongoing Validation Work
Ready to Partner?
Join ADVOCATE and leverage our Popperian AI architecture for your clinical trial.
Back to ADVOCATE Overview