# How to Be Oversurprised

post by royf · 2013-01-07T04:02:01.424Z · score: 13 (18 votes) · LW · GW · Legacy · 14 comments**Followup to:** How to Disentangle the Past and the Future

Some agents are memoryless, reacting to each new observation as it happens, without generating a persisting internal structure. When a LED observes voltage, it emits light, regardless of whether it did so a second earlier.

Other agents have *very* persistent memories. The internal structure of an amber nugget can remain unchanged by external conditions for millions of years.

Neither of these levels of memory persistence makes for very intelligent agents, because neither allows them to be good separators of the past and the future. Memoryless agents only have access to the most recent input of their sensors, which leaves them oblivious to the hidden internal structures of other things around them, and to the existence of things not around them. Unchanging agents, on the other hand, fail to entangle themselves with new evidence, which prevents them from keeping up to date with a changing world.

Intelligence requires observations. An intelligent agent needs to strike a delicate balance between the persistence of its internal structure and its susceptibility to new evidence. The optimal balance, a Bayesian update, has been explained many times before, and was shown to be optimal in keeping information about the world. This post highlights yet another aspect.

Suppose I predicted that a roll of a die will give 4, and then it did. How surprised would you be?

You may intuitively realize that the degree of your surprise should be a decreasing function of the probability that you assign to the event. Predicting a 4 on a fair die is more surprising than on one that is loaded in favor of 4. You may also want the measure of surprise to be extensive: if I repeated the feat a second time, you would be twice as surprised.

In that case, there's essentially one consistent notion of surprise (also called: self-information, surprisal). The amount of surprise that a random variable X has value x is

S(x) = -log Pr(X=x).

This is the negative logarithm of the probability of the event X=x.

This is a very useful concept in information theory. For example, the entropy of the random variable X is the surprise we expect to have upon seeing its value. The mutual information between X and another random variable Y is the difference between how much we expect x to surprise us now, and how much we expect it to surprise us after we first look at y.

The surprise of seeing 4 when rolling a fair die is

S(1/6) = -log(1/6) ≈ 2.585 bit.

That's also the surprise of any other result, so that's also the entropy of a die.

By the way, the objective surprise of a specific result is the same whether or not I actually announce it as a prediction. The reason you don't "feel surprised" when no prediction is made, is that your intuition evolved to successfully avoid hindsight bias. As we'll see in a moment, not all surprise is useful evidence.

The Bayesian update is optimal in that it gives the agent the largest possible gain in information about the world. Exactly how much information is gained?

We might ask this question when we choose between an agent that doesn't update at all and one that updates perfectly: this is what we stand to gain, and we can compare it with the cost (in energy, computation power, etc.) of actually performing the update.

We can also ask this question when we consider what observations to make. Not updating on new evidence is equivalent, in terms of information, to not gathering it in the first place. So the benefit of gathering some more evidence is, at most, the benefit of subsequently using it in a Bayesian update.

Suppose that at time t the world is in a state W_{t}, and that the agent may look at it and make an observation O_{t}. Objectively, the surprise of this observation would be

S_{obj} = S(O_{t}|W_{t}) = -log Pr(O_{t}|W_{t}).

However, the agent doesn't know the state of the world. Before seeing the observation at time t, the agent has its own memory state M_{t-1}, which is entangled with the state of the world through past observations, but is not enough for the agent to know everything about the world.

The agent has a *subjective* surprise upon seeing O_{t}, which is the result of its own private prior:

S_{subj} = S(O_{t}|M_{t-1}) = -log Pr(O_{t}|M_{t-1}),

and this may be significantly different than the objective surprise.

Interestingly, the amount of information that the agent stands to gain by making the new observation, and perfectly updating on it in a new memory state M_{t}, is exactly equal to the agent's expected *oversurprise* upon seeing the evidence. That is, any update from M_{t-1} to M_{t} gains at most this much information:

I(W_{t};M_{t}) - I(W_{t};M_{t-1}) ≤ E[S_{subj} - S_{obj}],

and this holds with equality when the update is Bayesian. (The math is below in a comment.)

For example, let's say I have two coins, one of them fair and the other turns up heads with probability 0.7. I don't know which coin is which, so I toss one of them at random. I expect it to show heads with probability

0.5 / 2 + 0.7 / 2 = 0.6,

so if it does I'll be surprised S(0.6) ≈ 0.737 bit, and if it doesn't I'll be surprised S(0.4) ≈ 1.322 bit. On average, I expect to be surprised

0.6 * S(0.6) + 0.4 * S(0.4) ≈ 0.971 bit.

The objective surprise depends on how the world really is. If I toss the fair coin, the objective surprise is S(0.5) = 1 bit for each of the two outcomes, and the expectation is 1 bit too. If I toss the loaded coin, the objective surprise is S(0.7) ≈ 0.515 bit for heads and S(0.3) ≈ 1.737 bit for tails, for an average of

0.7 * S(0.7) + 0.3 * S(0.3) ≈ 0.881 bit.

See how I'm a little undersurprised for the fair coin, but oversurprised for the loaded one. On average I'm oversurprised by

0.971 - (1 / 2 + 0.881 / 2) = 0.0305 bit.

By observing which side the coin turns up, I gain 0.971 bit of information, but most of it is just about this specific toss. Only 0.0305 bit of that information, my *over*surprise, goes beyond the specific toss to teach me something about the coin itself, if I update perfectly.

We are not interested in evidence for their own sake, only inasmuch as they teach us about the world. Winning the lottery is very surprising, but nobody gambles for epistemological reasons. When the odds are known in advance, the subjective surprise is exactly matched by the objective surprise - you are not oversurprised, and you learn nothing useful.

The oversurprise is the "spillover" of surprise beyond what is just about the observation, and onto the world. It's the part of the narrowness that originates in the world, not in the observation. And that's how much the observation teaches us about the world.

An interesting corollary is that you can never expect to be undersurprised, i.e. less surprised than you objectively should be. That's the same as saying that a Bayesian update can never lose information.

The next time you find yourself wondering how best to observe the world and gather information, ask this: how much more surprised do you expect to be, than someone who already knows what you wish to know.

**Continue reading:** Update Then Forget

## 14 comments

Comments sorted by top scores.

Suppose that at time t the world is in a state Wt, and that the agent may look at it and make an observation Ot. Objectively, the surprise of this observation would be Sobj = S(Ot|Wt) = -log Pr(Ot|Wt).

One note on philosophy of probability: if the world is in state Wt, what does it mean to say that an observation Ot has some probability given Wt? Surely all observations have probability 1 if the state of the world is exhaustively known.

Philosiphically, yes.

Practically, it may be useful to distinguish between a coin and a toss. The coin has persisting features which make it either fair or loaded for a long time, with correlation between past and future. The toss is transient, and essentially all information about it is lost when I put the coin away - except through the memory of agents.

So yes, the toss is a feature of the present state of the world. But it has the very special property, that given the bias of the coin, the toss is independent of the past and the future. It's sometimes more useful to treat a feature like that as an observation external to the world, but of course it "really" isn't.

Read Eliezer's http://lesswrong.com/lw/oj/probability_is_in_the_mind/ , I think it will answer your questions on this topic.

Thanks Sengachi. Actually I agree with that article, which is exactly why I mentioned the above. For an epistemically limited agent, it makes sense to talk about probabilities, because they are a measure of degree of belief.

But "Sobj" is supposed to be a measure of the "objective" surprisal you gain by making an observation Ot *given that you already know everything about the world's current state.* And it references probabilities. This, I suggest, makes no sense. Probability is *only* relevant for epistemically limited agents; if you are the Laplacian demon, you have no need to talk about probabilities at all; the past, present and future states of the world are a certainty for you.

Maybe it would be more accurate to replace S_obj with S_someone_who_already_knows_what_you_wish_to_know.

A human-like agent who knows which coin is fair and which is loaded (and is considering things at the level of coins and tosses) will have subjective probabilities like the 'objective' probabilities described in the article, while a more Laplacian demon-like agent who knows the positions and velocities of the atoms that compose the coins (and the air, and the arm of the tosser, etc.) will have subjective probabilities for outcomes corresponding to 'heads' and 'tails' that vary from toss to toss, and are much closer to 0 or 1 each time.

So then your over-surprise is relative to your ontology? Not sure what that implies for agents who can change their ontology...

Thanks for the post; I particularly enjoyed the one-sentence takeaway at the end. One criticism though: you use mathematical notation like I(W_t;M_t) without saying what it denotes. Even though that can be inferred from the surroundings, it would be less likely to confuse if you stated it explicitly.

I'm trying to balance between introducing terminology to new readers and not boring those who've read my previous posts. Thanks for the criticism, I'll use it (and its upvotes) to correct my balance.

When a LED observes voltage, it emits light, regardless of whether it did so a second earlier

I am not familiar with the details of LED lifespan, but a “white LED” in particular is not just an LED; it is a blue LED coated with down-converting phosphors to create a white color mix. Those phosphors have significantly shorter life, as noted in the article.

Also, *any* physical object will decay over time, so I'm not sure how good an example of a thing-which-can-change-state-but-has-no-memory can exist.

This one was fun for the math. Thank you. The practical advice is pretty prosaic - study the things you're most uncertain about.

Well, thank *you*!

Yes, I do this more for the math and the algorithms than for advice for humans.

Still, the advice is perhaps not so trivial: study *not* what you're most uncertain about (highest entropy given what you know) but those things with entropy generated by what you care about. And even this advice is incomplete - there's more to come.

Intelligence requires observations. An intelligent agent needs to strike a delicate balance between the persistence of its internal structure and its susceptibility to new evidence. The optimal balance, a Bayesian update, has been explained many times before, and was shown to be optimal in keeping information about the world.

What's optimal in terms of keeping information about the world is to store your entire sense data history - and everything you were born with. That is often *extremely* wasteful of memory (even if you compress) - but surely much more compact than keeping track of all the possible hypotheses that might explain your sense data, and their respective probabilities. I think it is hard to come up with a sense in which *that* is optimal.

You know, non-parametric methods along the lines of k-nearest neighbors basically do this. The cleverness in these methods is finding a log(n) access scheme for your data (a database problem, essentially, rather than a machine learning problem).

When the new memory state is generated by a Bayesian update from the previous one and the new observation , it's a sufficient statistic of these information sources for the world state , so that keeps all the information about the world that was remembered or observed:

=I(W_t;(M_{t-1},O_t)))

As this is all the information available, other ways to update can only have less information.

The amount of information gained by a Bayesian update is

)-I(W_t;M_{t-1}))

}{\Pr(W_t)\Pr(M_{t-1},O_t)}-\log\frac{\Pr(W_t,M_{t-1})}{\Pr(W_t)\Pr(M_{t-1})}\right])

}{\Pr(W_t,M_{t-1})}-\log\frac{\Pr(M_{t-1},O_t)}{\Pr(M_{t-1})}\right])

-\log\Pr(O_t%7CM_{t-1})])

and because the observation only depends on the world

+S(O_t%7CM_{t-1})])