User:IssaRice/Aumann's agreement theorem

From Machinelearning

Aumann's agreement theorem states that when two agents have the same priors and have common knowledge of each other's posteriors, then the posteriors must be the same.[1] The proof of the theorem is simple; almost all the difficulty of understanding the theorem is in understanding the formalism used and how it applies to real-life situations.

Information partitions

Actually, can't the information states overlap? What if, when Bob rolls a 1, Alice is told "1 or 3" and when Bob rolls a 2 is told "2 or 3" (with the rest of the situation remaining the same)? Then it seems like and , which overlap.

Similarly in Geanakoplos's example on p. 261,[2] in his example the agent is conveniently told whether the number is even or odd, giving a partition. But what if the agent is told whether the number is 1-7 or 2-8? It seems like this information states setup restricts the examples we can formalize, and the examples people consider seem to conveniently be formalizable.

In this case, if the real number happens to be 2-7, then the agent is told both "1-7" and "2-8", so he can deduce it is 2-7. Similarly, if he is told "1-7", then he can tell the number is 1, for if it wasn't 1, he would have also been told "2-8". So the information partition ends up being . So a rational agent can make their possible information states into a partition by reasoning about which clues are communicated. A general version of this is probably some basic theorem in game theory or a related field, but I am highly ignorant.

Actually, the above operation seems to be to "refine" the "partition" until it becomes an actual partition. The coarsest partition such that each element of the partition is a subset of the original set.

"which means we both know what the intersection is" [1] I don't think Aumann assumes we know which element of the other's partition we are in, so I don't think we know what the intersection is.

Join and meet

Given two partitions of a set, the join is the coarsest common refinement, and the meet is the finest common coarsening.

TODO give more formal definition

TODO give imperative definition [3]

TODO [2] gives a definition by defining at each

[4]

"So this “common knowledge” condition is tantamount to mind melding" [2]. But I think:

  • mind melding = join
  • common knowledge = meet

TODO aumann assumes the join of the partitions consists of nonempty events. it would be good to show where in his proof this assumption is used.

Common knowledge

Statement of theorem

Clarifying what each agent knows

  • the other's information partition?
  • the element of the other's information partition?
  • the intersection of the two elements of the information partitions?
  • the element of the meet of the information partitions?
  • the other's posterior?
  • the set of worlds where the other's posterior is equal to their current posterior?

If I'm Alice with information partition , then I can know without knowing what is, right? So we can build common knowledge and exchange posteriors without exchanging our exact information states.

Mechanics of reaching agreement

In the idealized agreement theorem setup, the process for two agents, Alice and Bob, agreeing on posteriors of some event will look like this:

  1. Alice and Bob each have an information partition.
  2. They each make some observation, updating their posteriors. At this point, their posteriors for might be the same, but might not.
  3. Alice and Bob tell each other their posteriors.
  4. After hearing the other's posterior, each modifies their own information partition. (The underlying sample space remains the same, but the partitioning changes, so that the posterior can change.)
  5. Repeat (1)–(4) until becomes common knowledge for some constants . In other words, keep exchanging posteriors until they become common knowledge. In step (2), after the first round the observation becomes the other's professed posterior.
  6. At this point, Aumann's agreement theorem applies, and we conclude . Alice and Bob agree!

What are the actual mechanics of observing the other's posterior and updating on it? I think what happens is that Bob gives Alice some set of omegas, and Alice gives Bob some other set of Omegas, and each intersects it with their current information set. And this resulting set becomes the new information set. Specifically suppose . Then Bob gives Alice the set . This way, Alice can carry out updates without knowing what is or even what is.

Maybe knowing and is equivalent to knowing the set above, in which case Alice can receive and and compute the set on her own.

Another confusion I have: Aumann's theorem only seems to say when the theorem applies at (6). But it seems like what we want to know in addition is that steps (1)–(5) converge, so that we can apply the theorem. Without knowing about convergence, it would seem like the theorem just doesn't apply in many cases. Aumann's paper concludes with "Our result implies that the process of exchanging information on the posteriors for will continue until these posteriors are equal." If the posteriors ever become common knowledge, then sure, but what if they never become common knowledge?

Given and , can we write in general? It seems like what we want is , but then . Is this okay? It seems like the convergence then happens in a single step? Maybe the examples I'm considering are too simple. Aumann's second coin toss example seems to have more than one exchange, so I should study that in detail. [3] uses the auxiliary sets to track updates rather than changing the information partitions.

There is a paper called something like "we can't disagree forever", which might be similar to the above. Ok, I just checked the paper, and it looks like this is exactly what the paper is about.

Hal Finney's example

This example is given by Hal Finney in a LessWrong comment.[5]

Let Alice and Bob be two agents. Each rolls a die, and knows what they rolled. In addition to this, each knows whether the other rolled something in the range 1–3 versus 4–6. As an example, suppose Alice rolls a 2 and Bob rolls a 3. Then Alice knows that the outcome is one of (2,1), (2,2), or (2,3), and Bob knows that the outcome is one of (1,3), (2,3), or (3,3).

Given the above description, what are the possible states of knowledge of Alice?

In the possible worlds notation, if we let be Alice's information partition and be Bob's information partition, we would write:

Now let be an arbitrary event, and let be constants. We now define the event

One of the assumptions in the agreement theorem is that is common knowledge. This seems like a pretty strange requirement, since it seems like the posterior probability of can never change no matter what else the agents condition on in addition to . For example, what if we bring in agent 3 and make the posteriors common knowledge again?

What if we take and say that agent 1 knows ?

In the form of above, we can change to be any subset of and to be any numbers in . We can also set the state of the world to be any . The agreement theorem says that as we vary these parameters, if we ever find that , then we must have .

Let be a random variable defined by . In other words, is the sum of the values of the two dice.

Explanation
(2, 3) 1 1 Given these parameters, so is common knowledge. This satisfies the requirement of the agreement theorem, and indeed 1=1.
(2, 3) 1/3 1/3 Given these parameters, so is common knowledge. This satisfies the requirement of the agreement theorem, and indeed 1/3=1/3.
(2, 3) 1/3 1/3 Given these parameters, , which is not a superset of , so is not common knowledge. Nonetheless, 1/3=1/3. (Is this a case of mutual knowledge that is not common knowledge?)
(2, 3) 0 1/3 Given these parameters, , which is not a superset of , so is not common knowledge.

Alice knows she rolled a 2 and Bob rolled something between 1 and 3. Now, consider that Alice is additionally told that Bob did not roll a 1. Now Alice's posterior probability of the event is 1/2. How does this affect the agreement theorem? It seems like Alice's information partition changes; in particular, Bob's roll gets divided into . Bob's information partition remains the same, so now . Now for the event , and and the set is not a superset of so the agreement theorem doesn't apply. And indeed, that's good because .

Another thing: can't Bob deduce that Alice must have rolled a 1 or 2 by knowing her posterior that X=4 is 1/2? If that's the case, it seems like the meet reduces to {(1,2), (1,3), (2,2), (2,3)}, which is the same as E, so the agreement theorem applies again, with the posterior of X=4 being 1/2. Actually, this might be the whole point of Aumanning, because otherwise the theorem just doesn't apply in most cases.

Aumann's first coin flip example

This example is from Aumann's original paper.[1]

let's hope i get this right. let A be the event "heads on the second toss", H be the event "heads on first toss" and T be the event "tails on first toss". B is the random variable representing the bias of the coin, and b is a specific value. B is distributed uniformly in .

and

Aumann's second coin flip example

This example is from Aumann's original paper.[1]

[6]

References