I'll build on my model of the previous post, and consider close and distant situation.

Recall that each world in is defined by , which is the number of people smiling in it. I'll extend the model by adding another Boolean variable , which determines, say, whether the world contains chocolate () or art (). So worlds can be described by the pair .

The default situation - the one if the AI does nothing - is , say. So smiling people in a world of art.

Then let's introduce two new partial preferences/pre-orders:

  • .
  • .

So say that, within a range of the default world (art, and the number of smiling people not being within people of ), the more smiling people, the better. While says that worlds in this range are better than worlds outside them.

These result in the following utility functions:

  • if , otherwise.
  • if , otherwise.

After normalisation, these become:

  • if , otherwise.
  • if , otherwise.

Again, I've felt free to translate to improve the clarity of the normalised version.

If we plot , we get the following:

Here I've slightly offset the (purple) from the (blue) worlds, for clarity of exposition, though they would of course be on top of each other.

Note that does not in itself distant situations, as this post recommends doing. Five close worlds are ranked about the distant worlds; but, conversely, five are close worlds are ranked below them.

To avoid distant situations, we need to add in , which explicitly punishes distant worlds, and hence plot :

This is much more like it! All the close worlds are now all ranked above the more distant ones.

But this is a close-run thing: the difference between the worse close world and the distant worlds is small. So, in general, when penalising distant worlds, we have to do this with some care, and maybe use a different standard of normalisation (since the downgrading of distant worlds is not really a preference of ours, rather a meta-tool of the synthesis method).

Frontpage
New Comment