Proofs Theorems 2,3

post by Diffractor · 2021-12-03T18:48:45.985Z · LW · GW · 0 comments

Theorem 2: Acausal Commutative Square: The following diagram commutes for any belief function . Any infradistribution  where  also makes this diagram commute and induces a belief function.

Here's how the proof proceeds. In order to show that the full diagram commutes, we first select properties on the four corners such that for each morphism, if the source fulfills the relevant property, then the target fulfills the relevant property. Obviously, the property for the bottom-right corner must be "is a belief function". There are eight phases for this, some of which are quite easy, but the two most difficult are at the start, corresponding to the two morphisms on the bottom side, because we must exhaustively verify the belief function conditions and infradistribution conditions. As usual, lower-semicontinuity is the trickiest part to show. Then, since we have four sides of the square to show are isomorphisms, we have another eight phases where we verify going forward to a different corner and back is identity. Finally, there's an easy step where we rule out that going around  in a loop produces a nontrivial automorphism, by showing that both paths from first-person static to third-person dynamic are equal.

First, let's define our properties.

The property for third-person static is , and  is an infradistribution.

The property for third-person dynamic is , and , and  is an infradistribution.

The property for first-person static is that  is a belief function.

The property for first-person dynamic is , and  when , and  when , and  is an infradistribution.

T2.1 Our first step is showing that all 8 morphisms induce the relevant property in the target if the source has the property.

T2.1.1 3-static to 1-static. Our definition is

and we must check the five conditions of a belief function for , assuming  is an infradistribution on  and it has the property 

T2.1.1.1 First, bounded Lipschitz constant. Our proof goal is

Let  be the Lipschitz constant of , which is finite because  is an infradistribution. Then let  be arbitrary (they must be bounded-continuous, though). We have

and then unpack the projection to get

and unpack the update to get

and then by Lipschitzness for ,

Now, when , the two functions are equal, so we have

We'll leave lower-semicontinuity for the end, it's hard.

T2.1.1.2 Now, normalization, ie  To do this, we have

Then this can be rewritten as

and then as a semidirect product.

and then, from our defining property for , and normalization for  since it's an infradistribution, we have

The same proof works for 0 as well.

T2.1.1.3 Now, sensible supports. Pick a  where  and  are identical on histories compatible with . We want to show

To start,

and undoing the projection and update, we have

Remember that actually,  is supported over , ie, the subset of  where the destiny is compatible with the policy, so when  and so  and  behave identically. Making this substitution in, we have

which then packs up as  and we're done.

T2.1.1.4 Now for agreement on max value. For the  type signature on inframeasures, we want to show

Let  be arbitrary. We have

and then the projection and update unpack to

and then, working in reverse from there, we get  and we're done. For the  type signature, we want to show

Let  be arbitrary. We have

which then unpacks as

and then, working in reverse, we get  and we're done.

T2.1.1.5 Time to revisit lower-semicontinuity. We want to show that

This is going to take some rather intricate work with Propositions 2 and 3, and Theorem 1 from Inframesures and Domain Theory [AF · GW]. Namely, the supremum of continuous functions is lower-semicontinuous, any lower-semicontinuous lower-bounded function has an ascending sequence of lower-bounded continuous functions which limits to it pointwise, and for any ascending sequence which limits to a lower-semicontinuous function pointwise, you can shuffle the limit outside of an inframeasure. Fix some continuous bounded , either in  or in , and sequence of policies  which converge to .

Our first task is to establish that the subset of  where  or  or ... or  is closed. This can be viewed as the following subset of :

 is compact, and  is compact as well, since it limits to a point and the limit point is included. So that product is a compact set, and must be closed. The latter set is  itself. And so, since this set is the intersection of a closed set with , it's closed in  with the subspace topology.

Now, define the functions  as follows. If  has  or  or  or... or , then return . Otherwise, return 1 (or  for the other type signature).

 is the function where, if , return , otherwise return 1 (or ).

Our first task is to show that all the , and  are lower-semicontinuous. This is pretty easy to do. They are 1 (or ) outside of a closed set, and less than that inside the closed set.

Lower-semicontinuity happens because we have one of three cases holding. In the first case, the limit point is outside of the closed set (in the complement open set). Then, at some finite stage and forever afterwards, the sequence is outside closed set of interest, and the sequence becomes just a constant value of 1 (or ), and so liminf equals the value of the limit.

In the second case, the limit point is inside the closed set, and there's a subsequence of policy-tagged destinies which remains within the closed set of interest. The liminf of the original sequence of policy-tagged destinies evaluated by  equals the liminf of the subsequence which stays in the "safe" closed set, because all other points in the sequence get maximal value according to  (because they're outside the closed set), so they're irrelevant for evaluating the liminf. Then, inside the closed set,  acts like , which is continuous, so the liminf of  inside the closed set equals the limit of  inside the closed set, which equals  applied to the limit point (in the closed set)

In the third case, the limit point is inside the closed set, but there's no subsequence of policy-tagged destinies which remains within the closed set of interest. So, after some finite stage, the value of  is 1 or infinity, so the liminf is high and the value of the limit point is low, and again we get lower-semicontinuity.

Finally, we need to show that  is an ascending sequence of functions which has  as the pointwise limit. The "ascending sequence" part is easy, because as m increases,  maps a smaller and smaller set (fewer policies) to the value  reports, and more points to 1 (or infinity).

If  has , then regardless of n, , so that works out. If  has , then the sequence  can't hit  infinitely often, otherwise , so after some finite n,  drops out of the closed set associated with , and we have .

Just one more piece of setup. For each , since it's lower-semicontinuous, we can apply our Proposition 3 from the domain theory post to construct an ascending sequence of bounded continuous functions  where  pointwise.

NOW we can begin showing lower-semicontinuity of .

by how  was defined. And then we unpack the projection and update to get

First, we observe that the function inside  is greater than , because  copies  when  or  or ... and is maximal otherwise, while the interior function just copies  when  and is maximal otherwise. So, by monotonicity, we can go

And then, we can use monotonicity again to get

Wait, what's this? Well, if , then , because  was constructed to be an ascending sequence of functions (in n) limiting to . And  is an ascending sequence of functions (in m), so it's below . Thus, the function  lies below , we just applied monotonicity.

Our next claim is that  is an ascending sequence of continuous functions which limits pointwise to . Here's how to show it. The functions are continuous because they're all the supremum of finitely many continuous functions . They ascend because

The first inequality was from considering one less function in the sup. the second inequality is because, for all m,  is an ascending sequence of functions in n.

For showing that the sequence  limits (in n) pointwise to , we fix an arbitrary  and do

The inequality is because eventually n goes past the fixed number , and then we're selecting a single function out of the supremum of functions. The equality is because  limits pointwise in  to . Now, with this, we can go:

Let's break that down. The first equality was because we showed that  is the pointwise limit of the . The second equality is because  makes an ascending sequence of functions, so the supremum of the functions would just be  itself. The first inequality is because , it's one of the approximating continuous functions. The second inequality was the thing we showed earlier, about how the limit exceeds  regardless of . Finally, we use again that  is the pointwise limit of the .

We just showed a quantity is greater than or equal to itself, so all these inequalities must be equalities, and we have

So, the ascending sequence of continuous functions  limits pointwise to . Now that we have that, let's recap what we've got so far.We have

Since  is an ascending sequence of continuous functions in , then by monotonicity of , that sequence must be increasing, so we have

and then, we can apply the monotone convergence theorem for infradistributions, that a sequence of continuous functions ascending pointwise to a lower-semicontinuous function has the limit of the infradistribution expectations equaling the expectation of the lower-semicontinuous supremum, so we have

and then we pack this back up as an update and projection

and back up to get

and lower-semicontinuity of  has been shown. So translating over an infradistribution makes a belief function.

T2.1.2 1-static to 3-static.

Now, we'll show the five infradistribution conditions on , along with the sixth condition on  that it equals .

First, to show it's even well-defined, we have

That inner function is lower-semicontinuous in , because, starting with

then we have that, because  is continuous and  is compact,  is uniformly continuous. So, as n increases, the function  uniformly limits to . Since all the  have a uniform upper bound on the Lipschitz constant (from  being a belief function) they converge to agreeing that  has similar value as , so we get

and then, by lower-semicontinuity for ,

So, since the inner function is lower-semicontinuous, it can be evaluated.

T2.1.2.1 Now for the condition that

We start with  and unpack the definition of , to get

And unpack the definition of semidirect product and what  means, to get

This step happens because, if , then  will be assessing the value of the constant-1 function (or constant-infinity function), and will return a maximal value (by condition 5 on belief functions, that they all agree on the value of 1 (or )). So the only way to minimize that line via a choice of  is to make it equal to , as that means that we're evaluating  instead, which may be less than the maximal value.

Now we pack up what  means, and the semidirect product, to get

Pack up the definition of  to get

Pack up the definition of "1-update on " to get

Pack up how projection mappings work, to get:

Pack up what  means and the semidirect product to get

So we have

as desired, since  was arbitrary.

T2.1.2.2 Now to verify the infradistribution conditions on . First, monotonicity. Let 

And unpack the definition of semidirect product and what  means, to get

Then, by monotonicity for all the , we have

which packs back up in reverse as .

T2.1.2.3 Now for concavity. Start with  and unpack in the usual way to get

Then, by concavity for the , as they're inframeasures, we have

and distribute the inf

Pulling the constants out, we have

and packing back up in the usual way, we have

Concavity is shown.

T2.1.2.4 For Lipschitzness, we start with  and then partially unpack in the usual way to get

And, since  is a sharp infradistribution, it has a Lipschitz constant of 1, so we have

Since all the have a uniform bound on their Lipschitz constants (one of the belief function conditions), we then have

which rearranges as

and we're done, we got a Lipschitz constant.

T2.1.2.5 The CAS property for  is trival because  is a closed subset of  which is a compact space, and so you can just take all of  as a compact support.

T2.1.2.6 That just leaves normalization. We have

by unpacking  and applying normalization for a belief function.

So now,  is an infradistribution.

T2.1.3 For 3-static to 3-dynamic, we have

 (which fulfills one of the 3-dynamic conditions), and 

Since  is an infradistribution,  is too, and for the condition on , we have:

and we're done, since  had the relevant condition.

T2.1.4 For 3-dynamic to 3-static, we have

First, we must show that the niceness conditions are fulfilled for the infinite semidirect product to be defined at all.  maps  to . This is clearly a continuous mapping, so in particular, it's lower-semicontinuous. Dirac-delta distributions are 1-Lipschitz when interpreted as inframeasures (all probability distributions have that property, actually). The compact-shared-CAS condition is redundant because  is already a compact space. And all probability distributions (dirac-delta distributions are probability distributions) map constants to the same constant, so we get the increase-in-constants property and the 1 maps to 1 property. So we can indeed take the infinite semidirect product of the .

Now, we'll show that unrolling like this just produces  exactly. We'll use  for the initial state (pair of policy and destiny, also written as ), and  for the successive unrolled history of states, actions, and observations. First, by unpacking 's definition,

Then we unpack the projection and semidirect product

And pack things up into a projection

And then we can observe that unrolling the initial state forever via "advance the destiny 1 step, popping actions and observations off, and advance the policy", when you project down to just the actions and observations (not the intermediate states), yields just the point distribution on that destiny, so we get:

And then substitute the value of that dirac-delta in to get

So, our . Now we can go

and we're done, showing  had the relevant condition from  having it. Since  is an infradistribution,  is too.

T2.1.5 For 3-dynamic to 1-dynamic, we have

For the action  (which is compatible with the start of the history), we have 

Applying our definition, we have

Unpacking the projection, we have (where  is a state)

Unpacking the update, we have

And then we consider that  is just the dirac-delta distribution on , so we have

Substitute the values in, and we have

and then pack up the dirac-delta and we get

So,  is just the dirac-delta distribution on  and the pair of  and the policy advanced one step, as it should be. For  that aren't  as it should be, we have

Applying our definition, we have

Unpacking the projection, we have (where  is a state)

Unpacking the update, we have

And then we consider that  is just the dirac-delta distribution on , so we have

Substitute the values in, and remembering that  and  are assumed to be different, and we have (regardless of )

Or infinity for the other type signature. Therefore, 

So our transition kernel works out. For the other conditions on , just observe that , and use the extreme similarity of their defining conditions.

T2.1.6 For 1-dynamic to 3-dynamic, we have  so we can use our defining property of  again, clearly show that  has the relevant defining properties, so that just leaves cleaning up the defining property for the  infrakernel. We remember that

Therefore, we have:

This occurs because the history must be compatible with the policy. Then, we can unpack as:

We unpack our , getting

Then, we observe that for any . So,  must be , and we have

Now,  is the dirac-delta distribution on , so making that substitution in, we have

Ie

And this holds for all , so we get our desired result that

T2.1.7 For 1-static to 1-dynamic, we just have

So this automatically makes the infrakernel have the desired properties. For showing the relevant properties of , we just copy the proof at the start of getting the properties for  from  being a belief function, since  and  are both defined in the same way.

T2.1.8 For 1-dynamic to 1-static, we have

We'll work on getting that latter quantity into a better form, but first we have to verify that
 is even well-defined at all, w need the  to fulfill the niceness conditions. They're defined as

For lower-semicontinuity,  is lower-semicontinuous, because it returns  on every action but one (this is a clopen set), and for the action that pairs up with the initial state, the policy-tagged history just has the history and policy advanced, which is continuous.  behaves similarly, continuous variation in input leads to continuous variation in output, because eventually the behavior of the policy settles down to a fixed action as the history of actions and observations stabilizes amongst one of the finitely many options. 

For making inframeasures,  always returns a dirac-delta distribution or , so we're good there. Similarly, for 1-Lipschitzness, both  and dirac-delta distributions are 1-Lipschitz. For compact-shared CAS, as the target space is , it's compact and you don't need to worry about that. Finally, both dirac-delta distributions and  map constants to either the same constant or higher, and in the  type signature, both  and dirac-delta distributions map 1 to 1. So all niceness conditions are fulfilled.

Let's work on unpacking the value

unpacking the projection and semidirect product, we get

And unpacking the update, we have

We can make a substitution which changes nothing (in the scope of the indicator function where ),  that

and then we can write this as a projection in the inside

And then we can realize that when we unroll , it just always deterministically unrolls the history, along with the policy, since for  to be an input state,  must be compatible with , so picking actions from  means there's never any action mismatches. Projecting this down to the actions and observations just makes the history  exactly. So we have

Substituting the dirac-delta value in, we have

And we pack the update and projection back up to get

Since  was arbitrary, we have

And so, an alternate way to define  is as

We can run through the same proof that  is indeed a belief function, back from the 3-static to 1-static case, because  fulfills all the same properties that  did.

Alright, now we have, for all four corners, a condition that's basically "came from an acausal belief function" that is preserved under morphisms. Now we need to show that all 8 back-and-forth directions are identity.

T2.2.1 For 1-static to 3-static back to 1-static, we want to show

This is (we assume the hard version first)

Then unpack the projection and update

Then unpack the semidirect product and  to get

Then realize that the minimizer is picking  exactly, otherwise you'd just get maximal value and all the  agree on what a maximal input maps to.

And we're done.

T.2.2.2 For 3-static to 1-static to 3-static being identity, we want to show

This is just exactly the condition we're assuming on , so we trivially fulfill this (the nontrivial part was shoved into showing that going belief-function to infradistribution over policy-tagged states produced this property).

T2.2.3 For 3-static to 3-dynamic back to 3-static, we need to show that

For this, we abbreviate states beyond the first one as , so  is the initial state, and  is an element of . Taking the complicated part, it's

And then we unpack the semidirect product a bit, to get

And then we can write this as a projection, to get

Now, we observe that because , when repeatedly unrolled, is just emitting the actions and observations from the starting destiny/history, this projection is just the point distribution on the action-observation sequence that is h.

Then we evaluate the expectation, yielding

And we're done.

T2.2.4 For going 3-dynamic to 3-static to 3-dynamic, we need that

To show this, just reuse the exact same proof from above, just with  instead of . Also,  works out.

At this point, we've established isomorphism for two sides of the square. We've just got two more sides left to address, then showing that going around the square results in identity instead of a nontrivial automorphism.

T2.2.5,6 For translating back and forth between 3-dynamic and 1-dynamic, we observe that both translations keep the initial infradistribution over policy-tagged destinies the same, and our 3-dynamic to 1-dynamic, and 1-dynamic to 3-dynamic proofs verified that the  and  behave as they should when they go back and forth, so we don't need to worry about this.

T2.2.7 Next is 1-dynamic to 1-static to 1-dynamic. The infrakernel is guaranteed to have the right form, we just need to show that  is unchanged. So, we must show

But, since our condition on 1-dynamic (that we showed is a property of applying the morphism to go from any belief function to first-person dynamic) was

The only thing we need to show is that, regardless of ,

For if we could show that, then we could go:

and we'd be done. So, we'll show that 

instead, as that's all we need. Again, we'll massage the more complicated side until we get it into the simple form for the other side.

We undo the projection, and abbreviate states as , and action-observation-state sequences as , to yield

Then we reexpress the semidirect product, to yield

We unpack the initial state a fuzz to yield

We apply the update to yield

Then, we realize that if , we can swap  out for  in the relevant function associated with it.

Now, we can partially pack this up as a projection, to get

At this point, we can realize something interesting.  is "start with an initial policy of  and h compatible with , then repeatedly feed in actions as if they were created by  forever, and forget about the states, leaving just the action-observation sequence". Now, since the action (being produced by ) always lines up with what the  encoded in the state would do, this process never hits the  infradistribution, it keeps going on and on and advancing the history with no issue. In particular, the history unrolls to completion, and the resulting action-observation sequence you get would just be the original destiny packed up in the state. So, this infradistribution is just . Making this substitution in, we get

Now we pack up the update again, to get

And then realize that this is a projection, to get

And we're done, we showed our desired result to establish that going first-person dynamic to static back to dynamic is identity.

T2.2.8 Time for the last leg of the square, that going first-person static to first-person dynamic and first-person static is identity. We want to show that, regardless of ,

Again, like usual, we'll take the complicated thing and repeatedly rewrite it to reach the simpler thing, for an arbitrary function.

Using  for states, we can rewrite the projection, to get

Then we rewrite the semidirect product, and unpack the initial state into a policy  and destiny h compatible with , to get

=

We can then rewrite the interior as a projection to get

Now, it's time to rewrite the update. It rewrites as:

Now, since  inside the scope of the indicator function, we can rewrite as

And use our usual argument from before that  is just the dirac-delta distribution on h, to get

Now, we can unpack the semidirect product as well as what  means, to get:

Now, if , then that inner function turns into 1 (or infinity) which is assigned maximum value by all , so it's minimized when , so we get:

And we're done with the last "doing this path is identity" result. All that remains in our proof is just showing that taking both paths from one corner to another corner makes the same result, to show that going around in a loop is identity, ruling out nontrivial automorphisms.

T2.3 The easiest pair of corners for this is going from first-person static to third-person dynamic in both ways. Obviously, the transition kernel would be the same no matter which path you took, which just leaves verifying that the starting infradistribution is the same. Down one path, we have

Down the other path, we have

Obviously, both these paths produce the same result when you try to define  in both ways from . And we're done!

 

Theorem 3: Pseudocausal Commutative Square: The following diagram commutes for any choice of pseudocausal belief function . Any choice of infradistribution  where  also makes this diagram commute.

Again, the work we need to put in is getting conditions in the four corners that are "came from a pseudocausal belief function", then, for phase 1, verifying that all 8 morphisms preserve the relevant property in the target if the source had it, then for phase 2, verifying that all 8 ways of going from a corner to 1 away and back result in identity to get 4 isomorphisms, then finally showing that starting in one corner and going to the other corner by two different paths make the same result, to rule out nontrivial automorphisms.

The property for third-person static is 

The property for third-person dynamic is  and 

The property for first-person static is pseudocausality, 

The property for first-person dynamic is  and  when , and  when .

Time to show preservation of these properties by all the morphisms, as well as that the translations make a belief function/infradistribution.

Our first step is showing that all 8 morphisms induce the relevant property in the target if the source has the property.

T3.1.1 3-static to 1-static. Our definition is

and we must check the five conditions of a belief function, as well as pseudocausality.

T3.1.1.1 Checking the five conditions of a belief function proceeds in almost exactly the same way as checking the five conditions of a belief function in the "Acausal Commutative Square" theorem. We leave it as an exercise to the reader, and will just elaborate on the only nontrivial new argument needed.

The only nontrivial modification is in our proof of lower-semicontinuity. In order for it to work, we need that for any m, the subset of  where  or  or ... or  is closed. 

We can rephrase the set  as the projection down to the  coordinate of the set

 is compact, and  is compact as well, since it limits to a point and the limit point is included. So that product is a compact set. The set  is closed. So  is the intersection of a compact and a closed set, and so is compact. And projections of compact sets are compact, and compactness implies closure. So, the set of histories  is closed regardless of m, even if .

The rest of the proof works out with no issue, so translating over an infradistribution makes a belief function.

T3.1.1.2 We still have to check pseudocausality however. Our translation is

 

and we want to show

Reexpressing this desired statement in terms of , we have

Let  and  be arbitrary. Then the left side is

We unpack the update

And unpack the update again, to get

Looking at this, the function is  when h is compatible with both  and , and 1 otherwise. This is a greater function than  when h is compatible with just , and 1 otherwise, so by monotonicity for infradistributions, we can get:

and pack up the update to get

And we're done, we get pseudocausality.

T3.1.2 1-static to 3-static. Our translation is

We'll show the infradistribution conditions on . First, to show it's even well-defined, we have

That inner function is lower-semicontinuous in , because  was assumed to have that property in its first argument. Since the inner function is lower-semicontinuous, it can be evaluated.

Now for the condition that

We start with  and unpack the definition of , to get

And then undo the projection and semidirect product and what  means, to get

And then, by pseudocausality for  (for all  and also  because  is supported entirely on histories compatible with ), we can swap  out for

And then unpack the update, to get

Now we pack up what  means, and the semidirect product, to get

Write this as a projection

Pack up the definition of  to get

Pack up the definition of "1-update on " to get

Pack up what  means and the semidirect product and the projection to get

So we have

as desired, since  was arbitrary.

Now to verify the infradistribution conditions on . This proof is pretty much identical to the proof in the "Acausal Commutative Square" theorem, interested readers can fill it in.

T3.1.3 For 3-static to 3-dynamic, we have  (which fulfills one of the 3-dynamic conditions), and 

Since  is an infradistribution,  is too, and for the condition on , we have:

and we're done, since  had the relevant condition.

T3.1.4 For 3-dynamic to 3-static, we have

First, we must show that the niceness conditions are fulfilled for the infinite semidirect product to be defined.  maps  to . This is clearly a continuous mapping, so in particular, it's lower-semicontinuous. Dirac-delta distributions are 1-Lipschitz when interpreted as inframeasures (all probability distributions have that property, actually). The compact-shared-CAS condition is redundant because  is already a compact space. And all probability distributions (dirac-delta distributions are probability distributions) map constants to the same constant, so we get the increase-in-constants property and the 1 maps to 1 property. So we can indeed take the infinite semidirect product of the .

Now, we'll show that unrolling like this just produces  exactly. We'll use  for the initial destiny state, and  for the successive unrolled history of states, actions, and observations. First, by unpacking 's definition, 

Then we unpack the projection and semidirect product

And pack things up into a projection

And then we can observe that unrolling the initial destiny forever via "advance the destiny 1 step, popping actions and observations off", when you project down to just the actions and observations, yields just the point distribution on that destiny, so we get:

And then substitute the value of that dirac-delta in to get 

So, our . Now we can go

and we're done, showing  had the relevant condition from  having it. Since  is an infradistribution,  is too.

T3.1.5 For 3-dynamic to 1-dynamic, we have

Let's write  as  (unpacking the destiny a bit). Now, for the action  (which is compatible with the start of the history), we have 

Applying our definition, we have

Unpacking the projection, we have

Unpacking the update, we have

And then we consider that  is just the dirac-delta distribution on , so we have

Substitute the values in, and we have

So,  is just the dirac-delta distribution on  and , as it should be. 

For  that aren't  as it should be, we have

Applying our definition, we have

Unpacking the projection, we have

Unpacking the update, we have

And then we consider that  is just the dirac-delta distribution on , so we have

and because , substituting the dirac-delta in produces 1 (or infinity), ie, . Therefore, 

So our transition kernel works out. For the other conditions, just observe that , and use the extreme similarity of their defining conditions.

T3.1.6 For 1-dynamic to 3-dynamic, we have  so we can use our defining property to again, clearly show that the resulting 3-dynamic starting infradistribution has the relevant property, so that just leaves cleaning up the defining property for the  infrakernel. We remember that

Therefore, we have:

Then, we can unpack the semidirect product as:

We unpack our , getting

Then, we observe that for any . So,  must be , and we have

Now,  is the dirac-delta distribution on , so making that substitution in, we have 

And this holds for all , so we get our desired result that

T3.1.7: For 1-static to 1-dynamic, we just have

So this automatically makes the infrakernel have the desired properties. For showing the relevant property of , we just copy the proof at the start of getting the properties for  from  being a belief function, since  and  are both defined in the same way.

T3.1.8 For 1-dynamic to 1-static, we have

We'll work on getting that latter quantity into a better form, but first we have to verify that
 has the  fulfilling the niceness conditions. It's defined as

For lower-semicontinuity,  is lower-semicontinuous, because it returns  on every action but one (this is a clopen set), and for the action that pairs up with the initial destiny, it just advances the destiny, which is continuous.  behaves similarly, continuous variation in input leads to continuous variation in output, because eventually the behavior of the policy settles down to a fixed action as the history of actions and observations stabilizes amongst one of the finitely many options. 

For making inframeasures,  always returns a dirac-delta distribution or , so we're good there. Similarly, for 1-Lipschitzness, both  and dirac-delta distributions are 1-Lipschitz. For compact-shared CAS, as the target space is , it's compact and you don't need to worry about that. Finally, both dirac-delta distributions and  map constants to either the same constant or higher, and in the  type sinature, both  and dirac-delta distributions map 1 to 1. So all niceness conditions are fulfilled.

Let's work on unpacking the value

unpacking the projection and semidirect product, we get

And then we can write this as a projection in the inside

One of two things can happen. In the first case, h is compatible with , so playing  against it never hits , and it unrolls to completion, and projecting down just yields the dirac-delta distribution on h itself. In the second case, h is incompatible with , so eventually we hit , which maps everything to maximal value (either 1 or infinity). Thus, we can write this as an indicator function

Substituting the dirac-delta value in, we have

And we pack the update back up to get

Since  was arbitrary, we have

And so, an alternate way to define  is as

We can then run through the same proof that  is indeed a belief function, back from the 3-static to 1-static case, because  fulfills all the same properties that  did.

Alright, now we have, for all four corners, a condition that's basically "came from a pseudocausal belief function" that is preserved under morphisms. Now we need to show that all 8 back-and-forth directions are identity.

T3.2.1 For 1-static to 3-static back to 1-static, we want to show

This is (we assume the hard version first)

Then unpack the update

And the projection

Then unpack the semidirect product and  to get

then rewrite this as an update

And then we can apply pseudocausality of \BF to get that the minimizer must pick  exactly because  and so we get

and we're done.

T3.2.2 For 3-static to 1-static to 3-static, we must show

Well, the property for a third-person static  is exactly that.

T3.2.3 For 3-static to 3-dynamic to 3-static, we want that

Taking the complicated part, it's

We unpack the projection

And then we unpack the semidirect product a bit, to get

And then we can write this as a projection, to get

Now, we observe that because , when repeatedly unrolled, is just emitting the actions and observations from the starting destiny, this projection is just the point distribution on the action-observation sequence that is .

Then we evaluate the expectation, yielding

And we're done.

T3.2.4 For going 3-dynamic to 3-static to 3-dynamic, we need that

To show this, just reuse the exact same proof from above, just with  instead of . Also,  works out as it should.

At this point, we've established isomorphism for two sides of the square. We've just got two more sides left to address, then showing that going around the square results in identity instead of a nontrivial automorphism.

T3.2.5,6 For translating back and forth between 3-dynamic and 1-dynamic, we observe that both translations keep the initial infradistribution over destinies the same, and our 3-dynamic to 1-dynamic, and 1-dynamic to 3-dynamic proofs verified that the  and  behave as they should when they go back and forth, so we don't need to worry about this.

T3.2.7 Next is 1-dynamic to 1-static to 1-dynamic. The infrakernel is guaranteed to have the right form, we just need to show that  is unchanged. So, we must show

But, since our condition on 1-dynamic (that we showed is a property of applying the morphism to go from any belief function to first-person dynamic) was

The only thing we need to show is that, regardless of ,

For if we could show that, then we could go:

and we'd be done. However, from the 1-dynamic to 1-static proof of property preservation, we already showed that

back then, so we're just done.

T3.2.8 Time for the last leg of the square, that going first-person static to dynamic and back to static is identity. We want to show that, regardless of ,

To do this, we use the result from the 1-dynamic to 1-static proof of property preservation, that

Applying this fact with  as an abbreviation for , we can rewrite our proof goal equivalently with this fact as

But we showed this exact result when we showed that 1-static to 3-static to 1-static was identity, so we're done.

And we're done with the last "doing this path is identity" result. All that remains in our proof is just showing that taking both paths from one corner to another corner makes the same result, to show that going around in a loop is identity, ruling out nontrivial automorphisms.

T3.3 The easiest pair of corners for this is going from first-person static to third-person dynamic in both ways. Obviously, the transition kernel would be the same no matter which path you took, which just leaves verifying that the starting infradistribution is the same. Down one path, we have

Down the other path, we have

Obviously, both these paths produce the same result when you try to define  in both ways from . And we're done!

0 comments

Comments sorted by top scores.