Subjective Naturalism in Decision Theory: Savage vs. Jeffrey–Bolker

post by Daniel Herrmann (Whispermute), Aydin Mohseni (aydin-mohseni), ben_levinstein (benlev) · 2025-02-04T20:34:22.625Z · LW · GW · 22 comments

Contents

      Summary:
  1. Subjective Naturalism: Richness & Austerity
  1.1. Framework vs. Action-Guiding Rule
  2. Savage’s Framework
      2.1. The Rectangular Field and Its Problems
  3. The Jeffrey–Bolker Framework
    3.1. Basic Setup
    3.2. A Key Axiom (Informally) 
    3.3. Richness in Jeffrey-Bolker
    3.4. Austerity in Jeffrey-Bolker
  4. Comparison & Key Advantages
  5. Concluding Remarks
None
22 comments

Summary:

This post outlines how a view we call subjective naturalism[1] poses challenges to classical Savage-style decision theory. Subjective naturalism requires (i) richness (the ability to represent all propositions the agent can entertain, including self-referential ones) and (ii) austerity (excluding events the agent deems impossible). It is one way of making precise certain requirements of embedded agency [LW · GW]. We then present the Jeffrey–Bolker (JB) framework, which better accommodates an agent’s self-model and avoids forcing her to consider things she takes to be impossible.[2]

1. Subjective Naturalism: Richness & Austerity

A naturalistic perspective treats an agent as part of the physical world—just another system subject to the same laws. Among other constraints, we think this means:

  1. Richness: The model must include all the propositions the agent can meaningfully consider, including those about herself. If the agent can form a proposition “I will do X”, then that belongs in the space of propositions over which she has beliefs and (where appropriate) desirabilities.
  2. Austerity: The model should only include events the agent thinks are genuinely possible. If she is certain something cannot happen, the theory shouldn’t force her to rank or measure preferences for that scenario. (Formally, we can regard zero-probability events as excluded from the relevant algebra.)

A decision-theoretic framework that meets both conditions is subjectively naturalist: it reflects the agent’s own worldview fully (richness) but doesn’t outstrip that worldview (austerity).

1.1. Framework vs. Action-Guiding Rule

In the literature, “decision theory” can refer to (at least) two different kinds of things:

  1. A conceptual or mathematical framework for representing an agent’s beliefs, desires, and preferences. These frameworks show how one might encode uncertainty, evaluate outcomes, or measure utility—but they don’t necessarily dictate a unique rule for which choice to make in any given scenario.
  2. A decision rule or algorithmic procedure for taking an action—e.g., “choose the act that maximizes expected utility according to these probabilities,” or “choose so as to maximize causal expected utility.”

When people say “vNM decision theory”, “Savage’s decision theory”, or “Jeffrey–Bolker,” they sometimes shift back and forth between framework-level discussion (how to model an agent’s preferences and degrees of belief) and rule-level discussion (which choice is rational to make, given that model).

Recognizing this framework vs. decision rule distinction helps clarify how a single formalism (like Jeffrey–Bolker) can encode multiple theories of choice. We can thus separate the mathematical modeling of beliefs and utilities (“framework”) from the question of which choice is prescribed (“rule”). Here we are focusing on the choice of framework.

2. Savage’s Framework

Leonard Savage’s classic theory in The Foundations of Statistics (1954) organizes decision problems via:

The agent’s preference ordering ⪰ is defined over all possible acts . Under certain axioms—particularly the Sure-Thing Principle—Savage proves there exists a unique probability measure  on [6] and a bounded utility function  on  such that:

2.1. The Rectangular Field and Its Problems

A crucial step is the Rectangular Field Assumption: the agent’s preference ordering must extend over every function from  to . This often means considering acts like “if it rains, then a nuclear war occurs; if it does not rain, then aliens will attack,” even if the agent herself thinks that’s physically absurd.

With this in hand, we can see that from a subjectively naturalist standpoint Savage doesn't do well:

Thus, while Savage's theory is very useful for some purposes, it violates both conditions of subjective naturalism. 

3. The Jeffrey–Bolker Framework

Richard Jeffrey (The Logic of Decision, 1965) and Ethan Bolker (1967) introduced a different formal approach that addresses these worries.

3.1. Basic Setup

Instead of dividing the world into “states” and “acts,” JB theory starts with a Boolean algebra .[7] Each element  is a proposition the agent can meaningfully entertain. That includes not just “It rains” but also “I will pick up the pen,” “I will have credence  in  at time ” etc. Some of the core components are:

  1. A strictly positive probability measure  defined over .
  2. A desirability (or utility) function  (a signed measure) is also defined over .[8]
  3. The agent has a preference ordering  defined on .[9] Certain axioms—Averaging, Impartiality, and Continuity—ensure that  is representable by expected utility:

 

         with  iff .

3.2. A Key Axiom (Informally)
 

Here we consider the key axiom for Jeffrey-Bolker, just as an example so that people can get a flavour for the framework.[10] 

When averaging, plus another axiom (Impartiality[12]), and some structual/continuity conditions hold, a representation theorem (due to Bolker) shows that preference is captured by a unique–up-to–transformation [LW · GW] probability  and a signed measure , giving an expected utility structure.

3.3. Richness in Jeffrey-Bolker

In JB, the agent can have a proposition “I choose ” right in . That means the agent’s beliefs about herself—probabilities about her actions or mental states—fit seamlessly into her overall probability space. No artificial separation between “states” and “acts.”

Hence, richness is greatly improved: all relevant propositions live together in .

3.4. Austerity in Jeffrey-Bolker

Because  is just a Boolean algebra closed under logical operations, the agent isn’t forced to include bizarre “causal” connections she rules out as physically impossible. Bolker puts it bluntly:

“The ‘Bolker objection’ (which could just as well have been named the Jeffrey objection) says that it is unreasonable to ask a decision maker to express preferences about events or lotteries he feels cannot occur.”

(Bolker 1974, p. 80)

And Jeffrey notes:

“I take it to be the principal virtue of the present theory, that it makes no use of the notion of a gamble or of any other causal notion.”

(Jeffrey 1965, p. 157)

Thus, in JB theory, you can avoid the bizarre “If it rains, nuclear war” situation simply by never admitting that object[13] into the algebra. The algebra only includes propositions that the agent views as possible.

In this way, austerity is satisfied. The framework tracks the agent’s sense of what is possible and excludes everything else.

4. Comparison & Key Advantages

Here’s how Jeffrey–Bolker addresses typical critiques of Savage:[14]

Hence, from the viewpoint of subjective naturalism, JB theory neatly combines:

  1. Richness: The agent can represent her own actions in the same probability space that describes the rest of the world.
  2. Austerity: The agent excludes events that are truly impossible or have zero probability from her perspective—no forced ranking of propositions or situations that are inconsistent with the agent's perspective.

5. Concluding Remarks

In short, JB helps us take an embedded, subjectively naturalist view of the agent—one that is both richer and more austere in a mathematically coherent way. 

To be clear, we are not claiming that JB solves all problems of embedded or naturalized agency.[15] But we think it is a useful starting point, for the reasons above. 

  1. ^
  2. ^

    There are already discussions of these different frameworks on LessWrong. For example, Abram's discussion here [LW · GW]. This post is meant to complement such existing posts, and give our take on some of the conceptual differences between different decision theory frameworks.

  3. ^

    Savage's framework is similar to vNM, but superior in the sense that you don't assume the agent's degrees of belief obey the probability axioms, or even that she has degrees of belief in the first place. Rather, just as how in vNM we derive an agent's utility function from her preferences over gambles, in Savage we derive an agent's utilities and probabilities from her preferences over acts.

  4. ^

    EDT is often associated with the Jeffrey-Bolker framework since it is what Jeffrey initially wrote down in his framework, but the framework itself admits of different decision rules. 

  5. ^
  6. ^

    Really, we have a -algebra on  over which the probability measure is defined, and we can get integration, not just summation for expected value. To keep things more readable we'll stick with the states instead of the algebra over states, and we'll often write things down in sums instead of integration.

  7. ^

    This algebra is also complete and atomless:

    • An algebra is complete if every subset of the algebra has both a supremum and an infimum, relative to implication.
    • An algebra is atomless if for each member of the algebra  there is some other member of the algebra  other than the bottom element such that  implies .

    We also think that the atomlessness of the Jeffrey-Bolker algebra has a very naturalistic flavour, as partially spelled out by Kolmogorov, but we leave a more thorough discussion of this feature for a future post. 

  8. ^

    Technically it is not defined over the bottom element of the algebra.

  9. ^

    Again, minus the bottom element of the algebra.

  10. ^

    You can find a brief introduction to the full axioms here. We also really like this paper by Broome (1990).

  11. ^

    Averaging ensures that the disjunction of two propositions lies between the two propositions. For example, if you prefer visiting the Museum of Jurassic Technology to visiting the Getty Museum, then the prospect of visiting either the one or the other should be dispreferred to surely visiting the Museum of Jurassic Technology, and preferred to surely visiting the Getty Museum.

  12. ^

    A technical condition that effectively pinpoints when two disjoint propositions  and  are equiprobable, by checking how adding a third proposition  to each side does (or doesn’t) alter the preference.

  13. ^

    In decision theory, we often call the objects of preference "prospects". Thus we can think of the point here as noting that in JB, all prospects are propositions, whereas this isn't the case in something like Savage. 

  14. ^

    There are others, a bit beyond the scope of this post: 

  15. ^

22 comments

Comments sorted by top scores.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2025-02-05T01:15:43.528Z · LW(p) · GW(p)

Happy to see this. 

I have some very basic questions:

How can I see inconsistent preferences within the Jeffrey Bolker framework? What about incomplete preferences ?

Is there any relation you can smimagube with imprecise probability / infraprobability, i.e. knightian uncertainty ?

Replies from: Whispermute
comment by Daniel Herrmann (Whispermute) · 2025-02-05T11:48:51.644Z · LW(p) · GW(p)

The JB framework as standardly formulated assumes complete and consistent preferences. Of course, you can keep the same JB-style objects of preference (the propositions) and change modify the preference axioms. For incomplete preferences, there's a nice paper by Richard Bradley, Revising Incomplete Attitudes, that looks at incomplete attitudes in a very Jeffrey-Bolker style framework (all prospects are propositions). It has a nice discussion of different things that might lead to incompleteness (one of which is "Ignorance", related to the kind of Knightian uncertainty you asked about), and also some results and perspectives on attitude changes for imprecise Bayesian agents.

I'm less sure about inconsistent preferences - it depends what exactly you mean by that. Something related might be work on aggregating preferences, which can involve aggregating preferences that disagree and so look inconsistent. John Broome's paper Bolker-Jeffrey Expected Utility Theory and Axiomatic Utilitarianism is excellent on this - it examines both the technical foundations of JB and its connections to social choice and utilitarianism, proving a version of the Harsanyi Utilitarian Theorem in JB.

On imprecise probabilities: the JB framework actually has a built-in form of imprecision. Without additional constraints, the representation theorem gives non-unique probabilities (this is part of Bolker's uniqueness theorem). You can get uniqueness by adding extra conditions, like unbounded utility or primitive comparative probability judgments, but the basic framework allows for some probability imprecision. I'm not sure about deeper connections to infraprobability/Bayesianism, but given that these approaches often involve sets of probabilities, there may be interesting connections to explore.
 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2025-02-05T12:03:50.562Z · LW(p) · GW(p)

Mmmmm

Inconsistent and incomplete preferences are necessary for descriptive agent foundations. 

In vNM preference theory an inconsistent preference can be described as cyclic preferences that can be moneypumped. 

How to see this in JB ?

Replies from: Whispermute
comment by Daniel Herrmann (Whispermute) · 2025-02-05T12:05:25.035Z · LW(p) · GW(p)

Ah, so not like, A is strongly preferred to B and B is strongly preferred to A, but more of a violation of transitivity. Then I still think that the Broome paper is a place I'd look at, since you get that exact kind of structure in preference aggregation. 

The Bradley paper assumes everything is transitive throughout, so I don't think you get the kind of structure you want there. I'm not immediately aware of any work of that kind of inconsistency in JB that isn't in the social choice context, but there might be some. I'll take a look. 

There are ways to think about degrees and measures of incoherence, and how that connects up to decision making. I'm thinking mainly of this paper by Schervish, Seidenfeld, and Kadane, Measures of Incoherence: How Not to Gamble if You Must. There might a JB-style version of that kind of work, and if there isn't, I think it would be good to have one. 

But to your core goal or weakening the preference axioms to more realistic standards, you can definitely do that in JB by weakening the preference axioms, but still keeping the background objects of preference be propositions in a single algebra. I think this would still preserve many of what I consider the naturalistic advantages of the JB system. For modifying the preference axioms, I would guess descriptively you might want something like prospect theory, or something else along those broad lines. Also depends on what kinds of agents we want to describe. 
 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2025-02-07T21:51:35.037Z · LW(p) · GW(p)

I want to be able to describe agents that do not have (vNM, geometric, other) rational preferences because of incompleteness or inconsistency but self-modify to become so. 

Eg. In vNM utility theory there is a fairly natural weakening one can do which is ask for a vNM-style representation theorem after dropping transitivity.

[ Incidentally, there is some interesting math here having to do with conservative vs nonconservative vector fields and potentials theory all the way to hodge theory. ]

does JB support this ? 

Im confused since in vNM we start with a preference order over probability distributions. But in JB irs over propositions?

comment by Daniel Herrmann (Whispermute) · 2025-02-04T20:45:19.553Z · LW(p) · GW(p)

This post is partially motivated by a comment I made [LW · GW] on AnnaSalamon's [LW · GW] post about alternatives to VNM [LW · GW]. There I wanted to consider not just decision rule alternatives to vNM, but also decision framework alternatives to vNM. I hope that this post helps demonstrate that there can be value in thinking about the underlying frameworks we use. 

comment by quetzal_rainbow · 2025-02-05T06:07:23.516Z · LW(p) · GW(p)

I think austerity has a weird relationship with counterfactuals?

Replies from: Whispermute
comment by Daniel Herrmann (Whispermute) · 2025-02-05T11:18:14.575Z · LW(p) · GW(p)

Yes, austerity does have an interesting relationship with counterfactuals, which I personally consider a feature, not a bug. A strong version of austerity would rule out certain kinds of counterfactuals, particularly those that require considering events the agent is certain won't happen. This is because austerity requires us to only include events in our model that the agent considers genuinely possible.

However, this doesn't mean we can't in many cases make sense of apparently counterfactual reasoning. Often when we say things like "you should have done B instead of A" or "if I had chosen differently, I would have been richer", we're really making forward-looking claims about similar future situations rather than genuine counterfactuals about what could have happened.

For example, imagine a sequence of similar decision problems (similar as in, you view what you learn as one decision problem as informative about the others, in a straightforward way) where you must choose between rooms A and B (then A' and B', etc.), where one contains $100 and the other $0. After entering a room, you learn what was in both rooms before moving to the next choice. When we say "I made the wrong choice - I should have entered room B!" (for example, after learning that you chose the room with less money), from an austerity perspective we might reconstruct the useful part of this reasoning as not really making a claim about what could have happened. Instead, we're learning about the expected value of similar choices for future decisions, and considering the counterfactual is just an intuitive heuristic for doing that. If what was in room A is indicative of what will be in A', then this apparent counterfactual reasoning is actually forward-looking learning that informs future choices. Now of course not all uses of counterfactuals can get this kind of reconstruction, but at least many of them that seem useful can.

It's also worth noting that while austerity constrains counterfactuals, the JB framework can still accommodate causal decision theory approaches (like Joyce's or Bradley's versions) that many find attractive, and so in a sense allows certain kinds of decision-theoretic counterfactuals. Now, I think one could push back on austerity grounds even here, and I do think that some versions of CDT implemented in JB would run afoul of certain strong interpretations of austerity. However, I'd say that even with these additions, JB remains more austere than Savage's framework, which forces agents to rank clearly impossible acts.

The core insight is that we can capture much of the useful work done by counterfactual reasoning without violating austerity by reinterpreting apparently counterfactual claims as forward-looking learning opportunities.

Replies from: quetzal_rainbow
comment by quetzal_rainbow · 2025-02-07T08:32:29.182Z · LW(p) · GW(p)

I'm more worried about counterfactual mugging and transparent Newcomb. Am I right that you are saying "in first iteration of transparent Newcomb austere decision theory gets no more than 1000$ but then learns that if it modifies its decision theory into more UDT-like it will get more money in similar situations", turning it into something like son-of-CDT?

comment by Noosphere89 (sharmake-farah) · 2025-02-05T16:41:26.051Z · LW(p) · GW(p)
  1. Austerity: The model should only include events the agent thinks are genuinely possible. If she is certain something cannot happen, the theory shouldn’t force her to rank or measure preferences for that scenario. (Formally, we can regard zero-probability events as excluded from the relevant algebra.)

 

I'd want to mention that in infinite contexts, probability 0 events are still possible.

(An example here is possibly the constants of our universe, which currently are real numbers, but any specific real number has a 0 probability of being picked)

It's very important to recognize when you are in a domain such that probability 0 is not equal to impossible.

The dual case holds as well, that is probability 1 events are not equal to certainty in the general case.

Replies from: cubefox
comment by cubefox · 2025-02-06T10:55:07.310Z · LW(p) · GW(p)

If understand correctly, possible probability 0 events are ruled out for Kolmogorov's atomless system of probability mentioned in footnote 7

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2025-02-06T13:43:20.578Z · LW(p) · GW(p)

Wait, how does the atomless property ensure that if the probability of an event is 0, then the event can never happen at all, as a matter of logic?

Replies from: cubefox
comment by cubefox · 2025-02-07T15:37:17.092Z · LW(p) · GW(p)

The atomless property and only contradictions taking a 0 value could both be consequences of the axioms in question. The Kolmogorov paper (translated from French by Jeffrey) has the details, but from skimming it I don't immediately understand how it works.

comment by cubefox · 2025-02-14T20:07:55.327Z · LW(p) · GW(p)

Is there a particular reason to express utility frameworks with representation theorems, such as the one by Bolker? I assume one motivation for "representing" probabilities and utilities via preferences is the assumption, particularly in economics, that preferences are more basic than beliefs and desires. However, representation arguments can be given in various directions, and no implication is made on which is more basic (which explains or "grounds" the others).

See the overview table of representation theorems here, and the remark beneath:

Notice that it is often possible to follow the arrows in circles—from preference to ordinal probability, from ordinal probability to cardinal probability, from cardinal probability and preference to expected utility, and from expected utility back to preference. Thus, although the arrows represent a mathematical relationship of representation, they do not represent a metaphysical relationship of grounding.

So rather than bothering with Bolker's numerous assumptions for his representation theorem, we could just take Jeffrey's desirability axiom:

If and then

Paired with the usual three probability axioms, the desirability axiom directly axiomatizes Jeffrey's utility theory, without going the path (detour?) of Bolker's representation theorem. We can also add as an axiom the plausible assumption (frequently used by Jeffrey) that

This lets us prove interesting formulas for operations like the utility of a negation (as derived by Jeffrey in his book) or the utility of an arbitrary non-exclusive disjunction (as I did it a while ago), analogous to the familiar formulas for probability, as well as providing a definition of conditional utility .

Note also that the tautology having 0 utility provides a zero point that makes utility a ratio scale, which means a utility function is not invariant under addition of arbitrary constants, which is stronger than what the usual representation theorems can enforce.

comment by Charlie Steiner · 2025-02-05T17:46:13.187Z · LW(p) · GW(p)

Would you agree that the Jeffrey-Bolker picture has stronger conditions? Rather than just needing the agent to tell you their preference ordering, they need to tell you a much more structured and theory-laden set of objects.

If you're interested in austerity it might be interesting to try to weaken the Jeffrey-Bolker requirements, or strengthen the Savage ones, to zoom in on what lets you get austerity.

Also, richness is possible in the Savage picture, you just have to stretch the definitions of "state," "action," and "consequence." In terms of the functional relationship, the action is just the thing the agent gives you a preference ordering over, and the state is just the stuff that, together with action, gives you a consequence, and the consequences are any set at all. The state doesn't have to be literally the state of the world, and the actions don't have to be discrete, external actions.

comment by Ape in the coat · 2025-02-05T06:57:08.061Z · LW(p) · GW(p)

Richness: The model must include all the propositions the agent can meaningfully consider, including those about herself. If the agent can form a proposition “I will do X”, then that belongs in the space of propositions over which she has beliefs and (where appropriate) desirabilities.

I see a potential problem here, depending on what exactly is meant by "can meaningfully consider".

Consider this set up:

You participate in the experiment for seven days. Every day you wake up in a room and can choose between two envelopes. One of them has 100$ the other is empty. Then your memory of this act is erased. At the end of the experiment you get all the money that you've managed to win.

On day one money are assigned to an envelope randomly. However, on all the next days the money are put in the envelope that you didn't pick on the previous day. You do not have any access to random number generators.

Is the model supposed to include credence for proposition "Today the money is in envelope 1" when you wake up participating in such experiment?

Replies from: Whispermute
comment by Daniel Herrmann (Whispermute) · 2025-02-05T11:05:54.808Z · LW(p) · GW(p)

Thanks for this example. I'm not sure if I fully understand why this is supposed to pose a problem, but maybe it helps to say that by "meaningfully consider" we mean something like, is actually part of the agent's theory of the world. In your situation, since the agent is considering which envelope to take, I would guess that to satisfy richness she should have a credence in the proposition. 

I think (maybe?) what makes this case tricky or counterintuitive is that the agent seems to lack any basis for forming beliefs about which envelope contains the money - their memory is erased each time and the location depends on their previous (now forgotten) choice.

However, this doesn't mean they can't or don't have credences about the envelope contents. From the agent's subjective perspective upon waking, they might assign 0.5 credence to each envelope containing the money, reasoning that they have no information to favor either envelope. Or they might have some other credence distribution based on their (perhaps incorrect) theory of how the experiment works.

The richness condition simply requires that if the agent does form such credences, they should be included in their algebra. We're not making claims about what propositions an agent should be able to form credences about, nor about whether those credences are well-calibrated. The framework aims to represent the agent's actual beliefs about the world, as in, how things are or might be from the agent's perspective, even in situations where forming accurate beliefs might be difficult or impossible.

This also connects to the austerity condition - if the agent truly believes it's impossible for them to have any credence about the envelope contents, then such propositions wouldn't be in their algebra. But that would be quite an unusual case, since most agents will form some beliefs in such situations, even if those beliefs end up being incorrect or poorly grounded.

Replies from: Ape in the coat
comment by Ape in the coat · 2025-02-05T12:13:06.659Z · LW(p) · GW(p)

I'm not sure if I fully understand why this is supposed to pose a problem, but maybe it helps to say that by "meaningfully consider" we mean something like, is actually part of the agent's theory of the world. In your situation, since the agent is considering which envelope to take, I would guess that to satisfy richness she should have a credence in the proposition. 

Okay, then I believe you definetely have a problem with this example and would be glad to show you where exactly.

I think (maybe?) what makes this case tricky or counterintuitive is that the agent seems to lack any basis for forming beliefs about which envelope contains the money - their memory is erased each time and the location depends on their previous (now forgotten) choice.

However, this doesn't mean they can't or don't have credences about the envelope contents. From the agent's subjective perspective upon waking, they might assign 0.5 credence to each envelope containing the money, reasoning that they have no information to favor either envelope.

Let's suppose that the agent does exactly that. Suppose they believe that on every awakening there is 50% chance that money is in envelope 1. Then picking envelope 1 every time will in expectation lead to winning 350$ per experiment.

But this is clearly false. The experiment is specifically designed in such a manner that the agent can win money only on the first awakening. On every other day (6 times out of 7) the money would be in the envelope 2. 

So should the agent believe that there is only 1/7 chance that money are in envelope 1 then? Also no. I suppose you can see why. As soon as he tries to act on such belief it will turn out that 6 times out of 7 the money are in envelope 1.

In fact, we can notice, that there is no coherent value of credence for statement "Today the money are in envelope 1" that would not lead the agent to irrational behavior. This is because the term "Today" is not well-defined in the setting of such experiment. 

By which I mean that in the same iteration of the experiment propositions including "Today" may not have a unique value. On the first day of the experiment statement "Today money are in envelope 1" may be true, while on the second day it may be false, so in the single iteration of the experiment that lasts 7 days the statement is simultaneously true and false!

Which means that "Today money are in envelope 1" isn't actually an event from the event space of the experiment and therefore doesn't have a probability value, as probability function's domain is event space.

But this is a nuance of formal probability theory that most people do not notice, or even try to ignore outright. Our intuitions are accustoimed to situations where statements about "Today" can be represented as well-defined events from the event space and therefore we assume that they can always be "meaningfully considered". 

And so if you try to base you decision theory framework on what feels as meaningfull to an agent instead of what is formalizable mathematically, you will end up with a bunch of paradoxical situations, like the one I've just described.

Replies from: Whispermute
comment by Daniel Herrmann (Whispermute) · 2025-02-05T12:33:47.214Z · LW(p) · GW(p)

Thanks for raising this important point. When modeling these situations carefully, we need to give terms like "today" a precise semantics that's well-defined for the agent. With proper semantics established, we can examine what credences make sense under different ways of handling indexicals. Matthias Hild's paper "Auto-epistemology and updating" demonstrates how to carefully construct time-indexed probability updates. We could then add centered worlds or other approaches for self-locating probabilities.

Some cases might lead to puzzles, particularly where epistemic fixed points don't exist. This might push us toward modeling credences differently or finding other solutions. But once we properly formalize "today" as an event, we can work on satisfying richness conditions. Whether this leads to inconsistent attitudes depends on what constraints we place on those attitudes - something that reasonable people might disagree about, as debates over sleeping beauty suggest.
 

Replies from: Ape in the coat
comment by Ape in the coat · 2025-02-05T15:59:01.693Z · LW(p) · GW(p)

There is, in fact, no way to formalize "Today" in a setting where the participant doesn't know which day it is, multiple days happens in the same iteration of probability experiment and probability estimate should be different on different days. Which the experiment I described demonstrates pretty well.

Framework of centered possible worlds is deeply flawed and completely unjustified. It's essentially talking about a different experiment instead of the stated one, or a different function instead of probability.

For your purposes, however it's not particularly important. All you need is to explicitly add the notion that propositions should be well-defined events. This will save you from all such paradoxical cases.

Replies from: jack-vandrunen
comment by Jack VanDrunen (jack-vandrunen) · 2025-02-05T18:22:26.043Z · LW(p) · GW(p)

You might be interested in this paper by Wolfgang Spohn on auto-epistemology and Sleeping Beauty (and related) problems (Sleeping Beauty starts on p. 388). Auto-epistemic models have more machinery than the basic model described in this post has, but I'm not sure there's anything special about your example that prevents it being modeled in a similar way.

Replies from: Ape in the coat
comment by Ape in the coat · 2025-02-06T06:52:02.016Z · LW(p) · GW(p)

Sleeping Beauty is more subtle problem, so it's less obvious why the application of centred possible worlds fails.

But in principle we can construct a similar argument. If we suppose that, in terms of the paper, ones epistemic state should follow function P', instead of P on awakening in Sleeping Beauty we get ourselves into this precarious situation:

P'(Today is Monday|Tails) = P'(Today is Tuesday|Tails) = 1/2

as this estimate stays true for both awakenings:

P'(At Least One Awakening Happens On Monday|Tails) = 1 -  P'(Today is Tuesday|Tails)^2 = 3/4 

While the actual credence should be 100%. Which gives an obvious opportunity to money pump the Beauty by bets on awakenings on the days in the experiment.

This problem, of course, doesn't happen when we simply keep using function P for which "Today is Monday" and "Today is Tuesday" are ill-defined, but instead:

P(Monday Awakening Happens in the Experiment|Tails) = 1

P(Tuesday Awakening Happens in the Experiment|Tails) = 1

and 

P(At Least One Awakening Happens On Monday|Tails) =  P(Monday Awakening Happens in the Experiment|Tails) = 1

But again, this is a more subtle situation. The initial example with money in envelope is superior in this regard, because it's immediately clear that there is no coherent value for P'(Money in Envelope 1) in the first place.