I grappled with Savage a few years ago, to re-establish the possibility of unbounded utility functions and a proper treatment of paradoxical games like St. Petersburg within a minimally modified version of his framework. It will take me a while to reload that material into my brain.
Meanwhile, your construction in the case of countable S depends on having only finite additivity, to construct a countable set of null sets whose union is not null. Finite vs. countable additivity was a live issue when Savage was writing, but my impression (as a non-professional in this area) is that finite additivity eventually fell out of favour. Too much becomes awkward or goes wrong, and non-measurable sets and the Banach-Tarski paradox are a small price to pay.
Is there still a counterexample to Strong Dominance in the setting of countable additivity, and taking S to be a measurable space with events limited to measurable subsets?
ETA: On countable additivity, Savage mentions that his construction might be carried out in that context also and cites Villegas has having done so (pp 43-44 of the 1972 edition; Villegas “On qualitative probability sigma-algebras,” Annals of Mathematical Statistics, 35, 1787-1796.).
Hi! Thank you so much for this thoughtful comment. It’s great connecting someone who has spent time deep in the weeds of Savage. It took me a long time to understand St. Petersburg, Banach-Tarski, and Villegas-Debreu: definitely not easy!
Loosely speaking, the tension here comes down to two ingredients implied by Savage's axioms:
I agree Countable Additivity (CA) is the workhorse of most fields, including empirical ML. The exceptions are decision theorists and philosophers, who remain enthusiastic about the assumption of the Universal Domain. I personally think preserving the Universal Domain is specifically important for the philosophy of AGI decision-making: an AGI, by definition, should be general and capable of evaluating all acts.
Finite Additivity (FA) is a consequence of the Universal Domain alongside other assumptions. Savage himself noted that his probability measures have to be FA. In our work, we further prove that Savage's probability measures must also be locally strongly finitely additive under the axiom of constructibility.
As you pointed out, Villegas (and Debreu) provide the axiomatic foundation of Subjective Expected Utility with Countable Additivity by abandoning the universal domain, and our impossbility does not apply.
To summarize the theoretical landscape of what is and isn't possible:
Thanks again for engaging so deeply!
A common coherence defense of EU is that it blocks money pumps and exploitation. Yet Savage's axioms usually make dominated acts tie some dominating acts in EU.
Epistemic status
Math claim precise; alignment implications speculative. The proofs are joint work with Jingni Yang; the framing here is mine. Full writeup here.
Why start with Savage, not vNM
Most coherence writing on LessWrong and the Alignment Forum targets vNM, which assumes a given probability measure. Savage's framework is more fundamental. It derives both utility and probability from preferences alone. If dominance fails here, the gap is upstream of vNM. The result below shows it does.
The claim
Let be the state space. Acts are functions from states to a nondegenerate real interval of consequences (i.e., an interval containing more than one point, such as ).
In plain English: under the axiom of constructibility, you can always construct an act that pays strictly more than a constant act across an entire positive-probability event, yet Savage's EU assigns them the same value. The improvement is real, state by state, but the representation cannot see it.
Proof idea
Savage's framework formally defines events as all arbitrary subsets of the state space (Savage 1954/1972, p. 21). His Postulates P5 and P6 together force the state space to be infinite (p. 39). Together, these imply Savage's EU [3] on the full event domain ( ), with a convex-ranged representing probability . Convex-ranged: for any event and any number , there is a subevent with .
Convex-rangedness implies that every singleton is null. If a singleton had positive probability, convex-rangedness would require a subset of with exactly half that probability, which is impossible.
In the countable case, is a countable union of null sets with , so the null sets cannot form a -ideal. By Armstrong's equivalence (1988), this forces local strong finite additivity (LSFA). In the uncountable case, set-theoretic work under the axiom of constructibility yields the same conclusion.
[4]
Either way, there exists an event and a partition such that
Since is monotonic, it has at most countably many discontinuities. Choose strictly below the upper bound of the consequence interval to be a right-continuity point of , and take a sequence so that:
Then define
and
We can bound the EU difference by discarding the first null partition pieces and overestimating the remainder:
because the first pieces are null, and on the tail the utility increment is at most .
Since , the right-hand side tends to . Hence
Dominance is violated.
What this means in practice
Not a money pump. No cyclic preferences. This failure is prior to any pump. Expected utility evaluation is completely blind to the difference between an act and a strictly better alternative on a positive-probability event. [5]
That violates the dominance property coherence was supposed to secure. Whether this indifference can be turned into exploitable menu behavior depends on further assumptions about compensation and trade -- but the core theorem stands independently of that question.
Simply put: is strictly better than on every state in a non-null event, yet . (Note that must take on infinitely many outcome values; Wakker (1993) proved strict dominance survives if restricted to simple, finitely-valued acts).
Nearest predecessor
Wakker (1993) proved that Savage's axioms usually imply violations of strict stochastic dominance, and Stinchcombe (1997) provided an example showing indifference between an act and one that pointwise dominates it for countably infinite states.
The dominance property here is more primitive than stochastic dominance, and the claim is stronger than a pure existence example. While Wakker and Stinchcombe provided specific constructions, I prove a structural impossibility theorem. Via a classical equivalence (Armstrong 1988), every Savage representing probability on the universal domain exhibits this pathology. The violation follows unconditionally for every Savage representation, not just a hand-picked prior.
Savage's framework necessarily generates these dominance failures. [6]
I suspect the universal domain does most of the work, but I have not been able to cleanly separate it from specifically Savagean structure such as P2 or P4.
Why the Savage setup matters
Whether the state space relevant to us is effectively infinite, and whether a coherence theorem for general agency should be formulated on Savage's full event domain or on a restricted event algebra, are questions I consider genuinely open. When philosophers invoke Savage's axioms, they rely on his idealized universal domain ( ). Without it, you cannot claim coherence dictates preferences over all possible strategies. This creates a dilemma.
Keep the universal domain, and you get the dominance failure proved above. Savage's own axioms, taken at face value, do not secure dominance.
Drop the universal domain to fit bounded computation, and you lose Savage's original universality. Savage wants all acts to have a measure, while the countably additive approach assumes only some "measurable acts" do.
Either way, the coherence pitch has a gap. The result does not claim any physical AI system needs . It claims the theoretical argument, "coherence implies EU, and EU means you can't be exploited," relies on a framework that breaks its own dominance property.
Possible repairs
These work, but require abandoning Savage's original universal-domain ambition, which is what underpins the strongest, most unconditional coherence claims.
Takeaway for alignment
I make the gap concrete. Savage's axioms on a universal domain admit strict pointwise dominance between acts of identical expected utility. I grant the axioms entirely and prove with perfect coherence, the representation does not secure statewise dominance, vindicating Thornley's warning from an alternative angle.
If the case for expected utility is that pressure toward coherence should drive agents toward exploitation-resistant choice, the conclusion does not follow. Shah identified a first gap; this theorem widens it. If this blindness persists into value learning, fitting an EU model to observed behavior may inherit the dominance gap, leaving inferred preferences unable to distinguish an act from a genuine statewise improvement. This raises the possibility that an agent whose EU representation carries this gap could, under some conditions, be steered into accepting dominated trades during sequential plan execution.
Concluding remarks
My result does not show that EU is wrong; I target Savage's universal-domain framework with subjective probabilities. The theorem shows that dominance violations follow inevitably from the axioms, not that rational agents should weakly prefer dominated acts. The precise claim:
The open question is whether any repair can close the dominance gap while preserving enough of Savage's universal domain ambition for the coherence argument to retain its philosophical force, or, whether every such repair sacrifices the universality that made the pitch compelling in the first place.
Appendix: Proof sketch for the uncountable case
The bridge from set theory to decision theory is Armstrong's equivalence. The null sets of a finitely additive probability on form a -ideal if and only if the measure is not locally strongly finitely additive.
To force a dominance failure, it suffices to show that a finitely additive probability on cannot have null sets forming a -ideal.
Countably infinite . Savage's axioms imply every singleton is null. If the null sets were a -ideal, then the countable union of all singletons, namely itself, would be null, contradicting .
Uncountable . Assume toward contradiction that a finitely additive probability on has -ideal null sets. Let be the additivity cardinal. One shows:
Once the null ideal fails to be a -ideal, Armstrong gives local strong finite additivity: there exists with partitioned into countably many null sets . This construction yields acts where dominates yet , violating dominance.
References
An event is a subset of the state space. An event is null if changes on that event never affect preference. Once probability is granted, a null event is simply a zero-probability event. ↩︎
The constructibility axiom is used in Wakker (1993) and Stinchcombe (1997). ↩︎
De Finetti and Savage both resisted countable additivity as a rationality constraint. Kadane, Schervish, and Seidenfeld (1999) give positive decision-theoretic reasons to take finite additivity seriously. ↩︎
See the Appendix for the full proof sketch of the uncountable case. ↩︎
This bears on Demski's posts on generalized Dutch-book arguments. If those arguments motivate EU representation, this result shows the further step to a dominance-respecting safety guarantee still does not follow. ↩︎
In fact, for every SEU representation, for every interior act , there exist infinitely many acts such that strictly dominates (or strictly dominates ) yet . The proof is the same: perturb by on the LSFA partition. ↩︎