In summary: A probabilistic version of the Payor's Lemma holds under the logic proposed in the Definability of Truth in Probabilistic Logic. This gives us modal fixed-point-esque group cooperation even under probabilistic guarantees.

EDIT 10/24: I think the way the way this post is framed is somewhat confused and should not be taken literally. However, I do stand by some of the core intuitions here.

Background

Payor's Lemma: If then

We assume two rules of inference:

  • Necessitation:
  • Distributivity:

Proof:

  1. by tautology;
  2. by 1 via necessitation and distributivity;
  3. , by assumption;
  4. from 2 and 3 by modus ponens;
  5. from 4 by necessitation;
  6. from 5 and 3 by modus ponens.

The Payor's Lemma is provable in all normal modal logics (as it can be proved in the weakest, because it only uses necessitation and distributivity). Its proof sidesteps the assertion of an arbitrary modal fixedpoint, does not require internal necessitation (), and provides the groundwork for Lobian handshake-based cooperation without Lob's theorem.

It is known that Lob's theorem fails to hold in reflective theories of logical uncertainty. However, a proof of a probabilistic Payor's lemma has been proposed, which modifies the rules of inference necessary to be:

  • Necessitation:
  • Weak Distributivity: where here we take to be an operator which returns True if the internal credence of is greater than and False if not. (Formalisms incoming).

The question is then: does there exist a consistent formalism under which these rules of inference hold? The answer is yes, and it is provided by Christiano 2012.

Setup

(Regurgitation and rewording of the relevant parts of the Definability of Truth)

Let be some language and be a theory over that language. Assume that is powerful enough to admit a Godel encoding and that it contains terms which correspond to the rational numbers Let be some fixed enumeration of all sentences in Let represent the Godel encoding of

We are interested in the existence and behavior of a function which assigns a real-valued probability in to each well-formed sentence of We are guaranteed the coherency of with the following assumptions:

  1. For all we have that
  2. For each tautology we have
  3. For each contradiction we have

Note: I think that 2 & 3 are redundant (as says John Baez), and that these axioms do not necessarily constrain to in and of themselves (hence the extra restriction). However, neither concern is relevant to the result.

A coherent corresponds to a distribution over models of A coherent which gives probability 1 to corresponds to a distribution over models of . We denote a function which generates a distribution over models of a given theory as

Syntactic-Probabilistic Correspondence: Observe that This allows us to interchange the notions of syntactic consequence and probabilistic certainty.

Now, we want to give sane probabilities to sentences which talk about the probability gives them. This means that we need some way of giving the ability to talk about itself.

Consider the formula takes as input the Godel encodings of sentences. contains arbitrarily precise information about In other words, if then the statement is True for all and the statement is False for all is fundamentally a part of the system, as opposed to being some metalanguage concept.

(These are identical properties to that represented in Christiano 2012 by I simply choose to represent with as it (1) reduces notational uncertainty and (2) seems to be more in the spirit of Godel's for provability logic).

Let denote the language created by affixing to Then, there exists a coherent for a given consistent theory over such that the following reflection principle is satisfied: In other words, implies

Proof

(From now, for simplicity, we use to refer to and to refer to You can think of this as fixing some theory and operating within it).

Let represent the sentence for some We abbreviate as Then, we have the following:

Probabilistic Payor's Lemma: If then

Proof as per Demski:

  1. by tautology;
  2. by 1 via weak distributivity,
  3. , by assumption;
  4. from 2 and 3 by modus ponens;
  5. from 4 by necessitation;
  6. from 5 and 3 by modus ponens.

Rules of Inference

Necessitation: If then by syntactic-probabilistic correspondence, so by the reflection principle we have and as such by syntactic-probabilistic correspondence.

Weak Distributivity: The proof of this is slightly more involved.

From we have (via correspondence) that so We want to prove that from this, or We can do casework on . If then weak distributivity follows from vacuousness. If then as so and therefore Then, is True by reflection, so by correspondence it follows that

(I'm pretty sure this modal logic, following necessitation and weak distributivity, is not normal (it's weaker than ). This may have some implications? But in the 'agent' context I don't think that restricting ourselves to modal logics makes sense).

Bots

Consider agents which return True to signify cooperation in a multi-agent Prisoner's Dilemma and False to signify defection. (Similar setup to Critch's ). Each agent has 'beliefs' representing their credences over all formal statements in their respective languages (we are assuming they share the same language: this is unnecessary).

Each agent has the ability to reason about their own 'beliefs' about the world arbitrarily precisely, and this allows them full knowledge of their utility function (if they are VNM agents, and up to the complexity of the world-states they can internally represent). Then, these agents can be modeled with Christiano's probabilistic logic! And I would argue it is natural to do so (you could easily imagine an agent having access to its own beliefs with arbitrary precision by, say, repeatedly querying its own preferences).

Then, if each behave in the following manner:

where and they will cooperate by the probabilistic Payor's lemma.

Proof:

  1. via conjunction;
  2. as if the -threshold is satisfied all others are as well;
  3. by probabilistic Payor.

This can be extended to arbitrarily many agents. Moreso, the valuable insight here is that cooperation is achieved when the evidence that the group cooperates exceeds each and every member's individual threshold for cooperation. A formalism of the intuitive strategy 'I will only cooperate if there are no defectors' (or perhaps 'we will only cooperate if there are no defectors').

It is important to note that any is going to be uncomputable. However, I think modeling agents as having arbitrary access to their beliefs is in line with existing 'ideal' models (think VNM -- I suspect that this formalism closely maps to VNM agents that have access to arbitrary information about their utility function, at least in the form of preferences), and these agents play well with modal fixedpoint cooperation.

Acknowledgements

This work was done while I was a 2023 Summer Research Fellow at the Center on Long-Term Risk. Many thanks to Abram Demski, my mentor who got me started on this project, as well as Sam Eisenstat for some helpful conversations. CLR was a great place to work! Would highly recommend if you're interested in s-risk reduction.

New Comment