Summary:
This post outlines how a view we call subjective naturalism[1] poses challenges to classical Savage-style decision theory. Subjective naturalism requires (i) richness (the ability to represent all propositions the agent can entertain, including self-referential ones) and (ii) austerity (excluding events the agent deems impossible). It is one way of making precise certain requirements of embedded agency. We then present the Jeffrey–Bolker (JB) framework, which better accommodates an agent’s self-model and avoids forcing her to consider things she takes to be impossible.[2]
A naturalistic perspective treats an agent as part of the physical world—just another system subject to the same laws. Among other constraints, we think this means:
A decision-theoretic framework that meets both conditions is subjectively naturalist: it reflects the agent’s own worldview fully (richness) but doesn’t outstrip that worldview (austerity).
In the literature, “decision theory” can refer to (at least) two different kinds of things:
When people say “vNM decision theory”, “Savage’s decision theory”, or “Jeffrey–Bolker,” they sometimes shift back and forth between framework-level discussion (how to model an agent’s preferences and degrees of belief) and rule-level discussion (which choice is rational to make, given that model).
Recognizing this framework vs. decision rule distinction helps clarify how a single formalism (like Jeffrey–Bolker) can encode multiple theories of choice. We can thus separate the mathematical modeling of beliefs and utilities (“framework”) from the question of which choice is prescribed (“rule”). Here we are focusing on the choice of framework.
Leonard Savage’s classic theory in The Foundations of Statistics (1954) organizes decision problems via:
The agent’s preference ordering ⪰ is defined over all possible acts . Under certain axioms—particularly the Sure-Thing Principle—Savage proves there exists a unique probability measure on [6] and a bounded utility function on such that:
2.1. The Rectangular Field and Its Problems
A crucial step is the Rectangular Field Assumption: the agent’s preference ordering must extend over every function from to . This often means considering acts like “if it rains, then a nuclear war occurs; if it does not rain, then aliens will attack,” even if the agent herself thinks that’s physically absurd.
With this in hand, we can see that from a subjectively naturalist standpoint Savage doesn't do well:
Thus, while Savage's theory is very useful for some purposes, it violates both conditions of subjective naturalism.
Richard Jeffrey (The Logic of Decision, 1965) and Ethan Bolker (1967) introduced a different formal approach that addresses these worries.
Instead of dividing the world into “states” and “acts,” JB theory starts with a Boolean algebra .[7] Each element is a proposition the agent can meaningfully entertain. That includes not just “It rains” but also “I will pick up the pen,” “I will have credence in at time ” etc. Some of the core components are:
with iff .
Here we consider the key axiom for Jeffrey-Bolker, just as an example so that people can get a flavour for the framework.[10]
When averaging, plus another axiom (Impartiality[12]), and some structual/continuity conditions hold, a representation theorem (due to Bolker) shows that preference is captured by a unique–up-to–transformation probability and a signed measure , giving an expected utility structure.
In JB, the agent can have a proposition “I choose ” right in . That means the agent’s beliefs about herself—probabilities about her actions or mental states—fit seamlessly into her overall probability space. No artificial separation between “states” and “acts.”
Hence, richness is greatly improved: all relevant propositions live together in .
Because is just a Boolean algebra closed under logical operations, the agent isn’t forced to include bizarre “causal” connections she rules out as physically impossible. Bolker puts it bluntly:
“The ‘Bolker objection’ (which could just as well have been named the Jeffrey objection) says that it is unreasonable to ask a decision maker to express preferences about events or lotteries he feels cannot occur.”
And Jeffrey notes:
“I take it to be the principal virtue of the present theory, that it makes no use of the notion of a gamble or of any other causal notion.”
Thus, in JB theory, you can avoid the bizarre “If it rains, nuclear war” situation simply by never admitting that object[13] into the algebra. The algebra only includes propositions that the agent views as possible.
In this way, austerity is satisfied. The framework tracks the agent’s sense of what is possible and excludes everything else.
Here’s how Jeffrey–Bolker addresses typical critiques of Savage:[14]
Hence, from the viewpoint of subjective naturalism, JB theory neatly combines:
In short, JB helps us take an embedded, subjectively naturalist view of the agent—one that is both richer and more austere in a mathematically coherent way.
To be clear, we are not claiming that JB solves all problems of embedded or naturalized agency.[15] But we think it is a useful starting point, for the reasons above.
Following Daniel’s terminology.
There are already discussions of these different frameworks on LessWrong. For example, Abram's discussion here. This post is meant to complement such existing posts, and give our take on some of the conceptual differences between different decision theory frameworks.
Savage's framework is similar to vNM, but superior in the sense that you don't assume the agent's degrees of belief obey the probability axioms, or even that she has degrees of belief in the first place. Rather, just as how in vNM we derive an agent's utility function from her preferences over gambles, in Savage we derive an agent's utilities and probabilities from her preferences over acts.
EDT is often associated with the Jeffrey-Bolker framework since it is what Jeffrey initially wrote down in his framework, but the framework itself admits of different decision rules.
Really, we have a -algebra on over which the probability measure is defined, and we can get integration, not just summation for expected value. To keep things more readable we'll stick with the states instead of the algebra over states, and we'll often write things down in sums instead of integration.
This algebra is also complete and atomless:
We also think that the atomlessness of the Jeffrey-Bolker algebra has a very naturalistic flavour, as partially spelled out by Kolmogorov, but we leave a more thorough discussion of this feature for a future post.
Technically it is not defined over the bottom element of the algebra.
Again, minus the bottom element of the algebra.
You can find a brief introduction to the full axioms here. We also really like this paper by Broome (1990).
Averaging ensures that the disjunction of two propositions lies between the two propositions. For example, if you prefer visiting the Museum of Jurassic Technology to visiting the Getty Museum, then the prospect of visiting either the one or the other should be dispreferred to surely visiting the Museum of Jurassic Technology, and preferred to surely visiting the Getty Museum.
A technical condition that effectively pinpoints when two disjoint propositions and are equiprobable, by checking how adding a third proposition to each side does (or doesn’t) alter the preference.
In decision theory, we often call the objects of preference "prospects". Thus we can think of the point here as noting that in JB, all prospects are propositions, whereas this isn't the case in something like Savage.
There are others, a bit beyond the scope of this post: