Why medical AI calls itself Bayesian — and what it actually delivers
Thomas Bayes formulated that probability is not a frequency but a state of belief. P(A|B) = P(B|A) × P(A) / P(B). The prior, the likelihood, the posterior — an engine for updating knowledge.
In medicine, this philosophy is natural. A clinician performs intuitive Bayesianism: pre-test probability, test result, revised diagnosis.
Rare data (HDLSS), uncertainty quantification (distribution vs point score), integration of causal knowledge as directed graphs. The FDA and EMA recognise the value of this approach.
An arc in a Bayesian network does not say “A causes B.” It says “if I observe A, I must revise my belief about B.” Judea Pearl distinguishes three levels: observing, intervening, imagining. A Bayesian network operates at the first level. A causal graph at all three.
Maximum likelihood estimation, uniform or conjugate priors for convenience, uncertainty not propagated. The result looks Bayesian, sells as Bayesian, but lacks the properties.
A globally calibrated model can be catastrophically miscalibrated locally. Silent drift renders the model inadequate without alarm.
A senior data scientist mastering transformers costs €80–140k. A biostatistician mastering Bayesian networks costs €50k. The architectural choice is partly an HR choice.
Independent clinical validation. Honest uncertainty characterisation. Drift monitoring procedure. A team capable of maintaining the model over time.