Know what your model is actually doing — before you ship it.
Aletheia opens up foundation model internals to find shortcuts, confounders, and unstable reasoning — then packages what it finds into evidence your QA, regulatory, and ML teams can act on.
Pathology foundation models pass every benchmark thrown at them. But benchmarks don’t tell you whether the model is relying on gland architecture or staining darkness. That distinction matters when a new lab’s slides look different from training data.
Two pathways, one metric
Our experiments show models encode both a robust morphology pathway and a brittle stain-sensitive pathway. Standard AUC conflates them — you can’t tell how much of your 97% accuracy comes from biology and how much from color shortcuts.
Silent failure across labs
Staining protocols, scanner manufacturers, and tissue processing differ between hospitals. Features that correlate with tissue type at one site become noise at another. Distribution shift is invisible until a patient is misclassified.
Evidence that regulators can read
EU AI Act Article 13 requires transparency evidence for high-risk medical AI from August 2026. SHAP values and attention maps don’t answer the question regulators are asking: what concepts has this model learned, and are they sound?
Internal review is not neutral
The team that built the model is not well-positioned to audit its reasoning. Confirmation bias is structural. An independent interpretability audit provides evidence that stands up to external scrutiny.
How it works
From model internals to release-ready evidence.
Aletheia combines sparse autoencoder analysis with domain-specific stress tests and a structured evidence workflow. The output is a decision artifact, not a dashboard.
01
Intake & scoping
Define the safety claim, confirm model access, validate cohort breadth. Not every claim is testable — we figure that out before burning compute.
Semi-automated · 0.5–1 day
02
Representation extraction
Run validation cohorts through target layers and store the activations. These are the raw material for feature discovery — what the model has actually encoded.
Automated · Hours
03
Sparse feature discovery
Train sparse autoencoders on internal representations. Each learned feature maps to a single concept — a tissue type, a morphological pattern, or a staining artifact.
Automated · Hours–days
04
Biological triage
Classify every feature: is it biological signal or preparation artifact? Map clean features to tissue ontologies. Flag artifact-correlated features for causal testing.
Semi-automated · 1–2 days
05
Ablation & causal validation
Zero out specific features and measure the downstream effect. This separates features the model depends on from features that are just along for the ride.
Automated · 1–3 days
06
Evidence pack export
Structured report with feature atlas, ablation results, confounder register, cross-site stability analysis, and regulatory mapping. Human sign-off required.
Template-driven · <1 day
“The question is not whether your model is accurate. It’s whether it is accurate for the right reasons — and whether you can prove it when someone asks.”
Regulatory context
The window for evidence-light deployment is closing.
Three regulatory bodies are converging on the same requirement: AI medical devices must demonstrate that their internal reasoning is trustworthy — with documented evidence.
FDA — AI-Enabled Device Software Lifecycle Guidance
Draft guidance requires transparency and monitoring frameworks for AI/ML-based software as a medical device across the entire product lifecycle.
2026
EU AI Act — High-risk AI system obligations
Medical device AI classified as high-risk must meet transparency, technical documentation, and human oversight requirements. Explainability is explicitly mandated under Article 13.
Active
MHRA — Project Glass Box (UK)
MHRA’s AI as a Medical Device programme targets explainability and transparency as core requirements for regulatory approval in Great Britain.
Get started
Book an introductory call. Know before you ship.
Design-partner engagements are open for pathology AI teams preparing for deployment or regulatory submission. Scoped, structured, delivered in days.
Prepared by Aletheia · For internal QA and regulatory use
Report ref
ATH-2026-0034
Model
Prov-GigaPath/NSCLC-v2
Date
28 Mar 2026
Status
DRAFT
1. Executive summary & release recommendation
Hold
The audit identified 38 confounder features across 312 total. The model demonstrates strong within-site performance but carries documented shortcut dependency that reduces cross-site robustness. Scanner artifact (feat_0203) confirmed causal via ablation (−0.14 AUC). Deployment to new sites not recommended without stain normalisation pipeline.
2. Feature atlas — 312 features catalogued
38 flagged
247 histological features mapped to NCIT, UBERON, and GO ontologies. 38 confounder features flagged. 27 features classified as ambiguous pending expert pathology review.
3. Causal validation — 312 ablation experiments
Methodology validated
Ablation protocol confirmed by 50 negative controls (max Δ = 0.04%). Confounder ablation improves cross-site accuracy by 5.8%.
4. Stress test results — 6 suites, 2 failed
2 failed
Cross-site generalisation and stain normalisation robustness tests failed. Scanner variation and subgroup fairness passed.
5. Regulatory alignment notes
Compliance ref
Report structured to support EU AI Act Article 13, FDA AIMD lifecycle guidance, and MHRA Project Glass Box requirements.
Release gate
Sign-off matrix & open actions
Current gate status
Hold
Blocked by scanner shortcut and cross-site stability failure. 3 sign-offs pending.
3
Pending sign-offs
2
Open blockers
V1
Evidence pack version
Sign-off matrix
ML Lead
Needs replay data on remediation branch before sign-off.
pending
Validation / QA
Requires site-specific limitation statement before committee review.
pending
Regulatory partner
Method appendix acceptable. Decision depends on final remediation wording.
review
Founder / Product owner
Executive summary drafted. Ready to finalise once blockers resolve.
ready
Open actions before release
Complete replay on remediation branch — verify site-holdout delta improves after stain normalisation.
Draft site-specific limitation statement — QA requires bounded limitation text for external sites.
Update executive summary — reflect final remediation status and limitation wording.