Intelligent Agent Foundations Forumsign up / log in
Intertheoretic utility comparison: simple theory
discussion post by Stuart Armstrong 147 days ago | Jessica Taylor likes this | 8 comments

There’s been a lot of work on how to reach agreement between people with different preferences or values. In practice, reaching agreement can be tricky, because of issues of extortion/trade and how the negotiations actually play out.\(\newcommand{\S}{\mathbb{S}}\)\(\newcommand{\R}{\mathbb{R}}\)\(\newcommand{\U}{\mathbb{U}}\)\(\newcommand{\O}{\mathbb{O}}\)\(\newcommand{\H}{\mathcal{H}}\)

To put those issues aside, let’s consider a much simpler case: where a single agent is uncertain about their own utility function. Then there is no issue of extortion, because the agent’s opponent is simply itself.

This type of comparison is called intertheoretic, rather than interpersonal.

A question of scale

It would seem that if the agent believed with probability \(p\) that it followed utility \(u\), and \(1-p\) that it followed utility \(v\), then it should simply follow utility \(w=(p)u + (1-p)v\).

But this is problematic, because \(u\) and \(v\) are only defined up to positive affine transformations. Translations are not a problem: sending \(u\) to \(u+c\) sends \(w\) to \(w+pc\). But scalings are: sending \(u\) to \(ru\) does not usually send \(w\) to any scaled version of \(w\).

So if we identify \([u]\) as the equivalence class of utilities equivalent to \(u\), then we can write \(w=(p)u + (1-p)v\), but it’s not meaningful to write \([w]=(p)[u] + (1-p)[v]\).

From clarity, we’ll call things like \(u\) (which map worlds to real values) utility functions, while \([u]\) will be called utility classes.

The setup

This is work done in collaboration with Toby Ord, Owen Cotton-Barratt, and Will MacAskill. We had some slightly different emphases during that process. In this post, I’ll present my preferred version, while adding the more general approach at the end.

We will need the structure described in this post:

  1. A finite set \(\mathbb{S}\) of deterministic strategies the agent can take.
  2. A set \(\mathbb{U}\) of utility classes the agent might follow.
  3. A distribution \(p\) over \(\mathbb{U}\), reflecting the agent’s uncertainty over its own utility functions.
  4. Let \(\mathbb{U}_p \subset \mathbb{U}\) be the subset to which \(p\) assigns a non-zero weight. We’ll assume \(p\) puts no weight on trivial, constant utility functions.

We’ll assume here that \(p\) never gets updated, that the agent never sees any evidence that changes its values. The issue of updating \(p\) is analysed in the sections on reward learning agent.

We’ll be assuming that there is some function \(f\) that takes in \(\mathbb{S}\) and \(p\) and outputs a single utility class \(f(\mathbb{S},p) \in \mathbb{U}\) reflecting the agent’s values.

Basic axioms

  1. Relevant data: If the utility classes \([u]\) and \([v]\) have the same values on all of \(\mathbb{S}\), then they are interchangeable from \(f\)’s perspective. Thus, in the terminology of this post, we can identify \(\U\) with \(k(U)/\sim\).

This gives \(\U\) the structure of \(S \cup \{0\}\), where \(S\) is a sphere, and \(0\) corresponds to the trivial utility that is equal on all \(\S\). The topology of \(\U\) is the standard topology on \(S\), and the only open set containing \(\{0\}\) is the whole of \(\U\).

Then with a reasonable topology on the probability distribution on \(\U\) – such as the weak topology? – this leads to the next axiom:

  1. Continuity: the function \(f\) is continuous in \(p\).

  2. Individual normalisation: there is a function \(h\) that maps \(\mathbb{U}\) to individual utility functions, such that \(f(\mathbb{S},p)= \left[\int_{\mathbb{U}_p} h(u)p \right]\) (using \(p\) as a measure on \(\mathbb{U}_p\)).

The previous axiom means that all utility classes get normalised individually, then added together according to their weight in \(p\).

  1. Symmetry: If \(\rho\) is a stable permutation of \(\mathbb{S}\), then \(f(\mathbb{S},p) \circ \rho = f(\mathbb{S},p \circ \rho)\).

Symmetry essentially means that the labels of \(\mathbb{S}\), or the details of how the strategies are implemented, do not matter.

  1. Utility reflection: \(h[u]=-h[-u]\).

  2. Cloning indifference: If there exists \(s_1, s_2 \in\mathbb{S}\) such that for all \(u\) in \(\mathbb{U}\) on which \(p\) is non-zero, \(u(s_1)=u(s_2)\), then \(f(\mathbb{S},p)=f(\mathbb{S}-\{s_1\},p)\).

Cloning indifference means that the normalisation procedure does not care about multiple strategies that are equivalent on all possible utilities: it treats these strategies as if they were a single strategy.

We might want a stronger result, an independence of irrelevant alternatives. But this clashes with symmetry, so the following axioms attempt to get a weaker version of that requirement.

Relevance axioms

The above axioms are sufficient for the basics, but, as we’ll see, they’re compatible with a lot of different ways of combining utilities. The following two axioms attempt to put some sort of limitations on these possibilities.

First of all, we want to define events that are irrelevant. In the terminology of this post, let \(ha\) be a partial history (ending in an action), with at two possible observations afterwards: \(o\) and \(o'\).

Then \(\S_{ha}=\S_{hao}\times\S_{hao'}\). Then if there exists a bijection \(\sigma\) between \(\S_{hao}\) and \(\S_{hao '}\) such that, for all \(u\) with \([u]\in\mathbb{U}_p\), \(u(s)=u(\sigma(s))\), then the observation \(o\) versus \(o'\) is irrelevant. See here for more on how to define \(u\) on \(\S_{hao}\) in this context.

Thus irrelevance means that the utilities in \(\mathbb{U}_p\) really do not ‘care’ about \(o\) versus \(o'\), and that the increased strategy set it allows is specious. So if we remove \(o\) as a possible observation (substituting \(o'\) instead) this should make no difference:

  1. Weak irrelevance: If \(o\) versus \(o'\) given \(ha\) is irrelevant for \(p\), then making \(o\) (xor \(o'\)) impossible does not change \(f\).

  2. Strong irrelevance: If \(o\) versus \(o'\) given \(h\) is irrelevant for \(p\) and there is at least one other possible observation \(o''\) after \(ha\), then making \(o\) (xor \(o'\)) impossible does not change \(f\).

Full theory

In our full analysis, we considered other approaches and properties, and I’ll briefly list them here.

First of all, there is a set of prospects/options \(\mathbb{O}\) that may be different from the set of strategies \(\mathbb{S}\). This allows you to add other moral considerations, not just strictly consequentialist expected utility reasoning.

In this context, the \(f\) defined above was called a ‘rating function’, that rated the various utilities. With \(\mathbb{O}\), there are two other possibilities, the ‘choice function’ which selected the best option, and the permissibility function, which lists the options you are allowed to take.

If we’re considering options as outputs, rather than utilities, then we can do things like requiring the options to be Pareto only. We could also consider that the normalisation should stay the same if we remove the non-Pareto options or strategies. We might also consider that it’s the space of possible utilities that we should care about; so, for instance, if \(u(s_1)=1\), \(u(s_2)=0\) and \(u(s_3) = -1\), and similar results hold for all \([u]\) in \(\mathbb{U}_p\), then we may as well drop \(s_2\) from the strategy set as it’s image is in the mixture of the other strategies.

Finally, some of the axioms above were presented in weaker forms (eg the individual normalisations) or stronger (eg independence of irrelevant alternatives).

by Alex Mennen 146 days ago | link

You talk like \(p\) is countably supported, but everything you’ve said generalizes to arbitrary probability measures \(p\) over \(S\), if you replace “for all \(u\) assigned nonzero probability by \(p\)” with “for all \(u\) in some set assigned probability \(1\) by \(p\)”.

If you endow \(\mathbb{U}\) with the quotient topology from \(\mathbb{R^S}/\sim\), then the only open set containing \(0\) is all of \(\mathbb{U}\). This is a funny-looking topology, but I think it is ultimately the best one to use. With this topology, every function to \(\mathbb{U}\) is continuous at any point that maps to \(0\). As a consequence, the assumption “if \(f(\mathbb{S},p)\in S\)” in the continuity axiom is unnecessary. More importantly, what topology on the space of probability distributions did you have in mind? Probably the weak topology?

I find independence of irrelevant alternatives more compelling than symmetry, but as long as we’re accepting symmetry instead, it probably makes sense to strengthen the assumption to isomorphism-invariance: If \(\rho:\mathbb{S}_1\rightarrow\mathbb{S}_2\) is a bijection, then \(f(\mathbb{S}_2,p)\circ\rho=f(\mathbb{S}_1,p\circ\rho)\).

The relevance axioms section is riddled with type errors. \(u(s)=u(\sigma(s))\) only makes sense if \(\mathbb{S}=\mathbb{S}_{\neg X}\sqcup\mathbb{S}_{X}\), which would make sense if \(\mathbb{S}\) represented a space of outcomes rather than a space of strategies (which seems to me to be a more natural space to pay attention to anyway), or if \(X\) is fully under the agent’s control, whereas \(\mathbb{S}=\mathbb{S}_{\neg X}\times\mathbb{S}_{X}\) makes sense if \(X\) is fully observable to the agent. If \(X\) is neither fully under the agent’s control nor fully observable to the agent, then I don’t think either of these make sense. If we’re using \(\times\) instead of \(\sqcup\), then formalizing irrelevance seems trickier. The best I can come up with is that \(p\) is supported on \(u\) of the form \(u(s,t)=(1-q)\tilde{u}(\sigma(s))+q\tilde{u}(t)\), where \(q\) is the probability of \(X\). The weak and strong irrelevance axioms also contain type errors, since the types of the output and second input of \(f\) depend on its first input, though this can probably be fixed.

I didn’t understand any of the full theory section, so if any of that was important, it was too brief.


by Stuart Armstrong 146 days ago | link

Yes to your two initial points; I wanted to keep the exposition relatively simple.

Do you disagree with the reasoning presented in the picture-proof? That seems a simple argument against IIA. Isomorphism invariance makes sense, but I wanted to emphasise the inner structure of \(\mathbb{S}\).

Updated the irrelevance section to clarify that \(X\) is fully observed and happens before the agent takes any actions, and that \(u(s)\) should be read as \(u(s|\neg X)\).

The full theory section is to write up some old ideas, to show that the previous axioms are not set in stone but that other approaches are possible and were considered.


by Alex Mennen 146 days ago | link

Your picture proof looks correct, but it relies on symmetry, and I was saying that I prefer IIA instead of symmetry. I’m not particularly confident in my endorsement of IIA, but I am fairly confident in my non-endorsement of symmetry. In real situations, strategies/outcomes have a significant amount of internal structure which seems relevant and is not preserved by arbitrary permutations.

You’ve just replaced a type error with another type error. Elements of \(\mathbb{U}\) are just (equivalence classes of) functions \(\mathbb{S}\rightarrow\mathbb{R}\). Conditioning like that isn’t a supported operation.


by Stuart Armstrong 143 days ago | link

Ok, I chose the picture proof because it was a particularly simple example of symmetry. What kind of internal structure are you thinking of?


by Alex Mennen 142 days ago | link

For strategies: This ties back in to the situation where there’s an observable event \(X\) that you can condition your strategy on, and the strategy space has a product structure \(\mathbb{S}=\mathbb{S}_X\times\mathbb{S}_{\neg X}\). This product structure seems important, since you should generally expect utility functions \(u\) to factor in the sense that \(u(s,t)=qu_X(s)+(1-q)u_{\neg X}(t)\) for some functions \(u_X\) and \(u_{\neg X}\), where \(q\) is the probability of \(X\) (I think for the relevance section, you want to assume that whenever there is such a product structure, \(p\) is supported on utility functions that factor, and you can define conditional utility for such functions). Arbitrary permutations of \(\mathbb{S}\) that do not preserve the product structure don’t seem like true symmetries, and I don’t think it should be expected that an aggregation rule should be invariant under them. In the real world, there are many observations that people can and do take into account when deciding what to do, so a good model of strategy-space should have a very rich structure.

For outcomes, which is what utility functions should be defined on anyway: Outcomes differ in terms of how achievable they are. I have an intuition that if an outcome is impossible, then removing it from the model shouldn’t have much effect. Like, you shouldn’t be able to rig the aggregator function in favor of moral theory 1 as opposed to moral theory 2 by having the model take into account all the possible outcomes that could realistically be achieved, and also a bunch of impossible outcomes that theory 2 thinks are either really good or really bad, and theory 1 thinks are close to neutral. A natural counter-argument is that before you know which outcomes are impossible, any Pareto-optimal way of aggregating your possible preference functions must not change based on what turns out to be achievable; I’ll have to think about that more. Also, approximate symmetries between peoples’ preferences seem relevant to interpersonal utility comparison in practice, in the sense that two peoples’ preferences tend to look fairly similar to each other in structure, but with each person’s utility function centered largely around what happens to themselves instead of the other person, and this seems to help us make comparisons of the form “the difference between outcomes 1 and 2 is more important for person A than for person B”; I’m not sure if this way of describing it is making sense.


by Stuart Armstrong 141 days ago | link

I think I’ve got something that works; I’ll post it tomorrow.


by Stuart Armstrong 141 days ago | link

OK, got a better formalism:


by Stuart Armstrong 143 days ago | link

You’re right. I’ve drawn the set of utility functions too broadly. I’ll attempt to fix this in the post.






Note that the problem with
by Vadim Kosoy on Open Problems Regarding Counterfactuals: An Introd... | 0 likes

Typos on page 5: *
by Vadim Kosoy on Open Problems Regarding Counterfactuals: An Introd... | 0 likes

Ah, you're right. So gain
by Abram Demski on Smoking Lesion Steelman | 0 likes

> Do you have ideas for how
by Jessica Taylor on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

I think I understand what
by Wei Dai on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

>You don’t have to solve
by Wei Dai on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

Your confusion is because you
by Vadim Kosoy on Delegative Inverse Reinforcement Learning | 0 likes

My confusion is the
by Tom Everitt on Delegative Inverse Reinforcement Learning | 0 likes

> First of all, it seems to
by Abram Demski on Smoking Lesion Steelman | 0 likes

> figure out what my values
by Vladimir Slepnev on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

I agree that selection bias
by Jessica Taylor on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

>It seems quite plausible
by Wei Dai on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

> defending against this type
by Paul Christiano on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

2. I think that we can avoid
by Paul Christiano on Autopoietic systems and difficulty of AGI alignmen... | 0 likes

I hope you stay engaged with
by Wei Dai on Autopoietic systems and difficulty of AGI alignmen... | 0 likes


Privacy & Terms