Intelligent Agent Foundations Forumsign up / log in
Acausal trade: introduction
post by Stuart Armstrong 580 days ago | discuss

A putative new idea for AI control; index here.

I’ve never really understood acausal trade. So in a short series of posts, I’ll attempt to analyse the concept sufficiently that I can grasp it - and hopefully so others can grasp it as well.


Other posts in the series: Introduction, Double decrease, Pre-existence deals, Full decision algorithms, Breaking acausal trade, Trade in different types of utility functions, Being unusual, and Summary.

The simplest model

There are \(N\) different rooms, with potential agents in them. The probability of the agents existing is a distribution \(Q\), with marginal probabilities \(q_i\), representing the probability that agent \(A_i\) exists in room \(i\). That agent has a utility \(u_i\), which they are motivated to maximise.

The agents will never meet, never interact in any way, won’t even be sure of each other’s existence, may not known \(N\), and may have uncertainty over the values of the other \(u_j\)’s. Each agent only acts in their own room. They may choose to diminish \(u_i\) to increase one or more other \(u_j\) with \(i\neq j\); this is what allows the possibility of trade.

Infinities, utility weights, negotiations, trade before existence

There are a number of things I won’t be considering here. First of all, infinities. In reality, acausal trade would happen in the real universe, which is likely infinite. It’s not clear at all how to rank infinitely many causally disconnected world-pieces. So I’ll avoid that entirely, assuming \(N\) is finite (though possibly large).

There’s also the thorny issue of how to weigh and compare different utility functions, and/or the process of negotiation about how to divide the gains from trade.

I’ll ignore all these issues, and see the \(u_i\) as functions from states of the world to real numbers: individual representatives of utility functions, not equivalence classes of equivalence functions. And the bargaining will be a straight one for one increase and decrease: a fair deal is one where \(u_i\) and \(u_j\) get the same benefit – as measured by \(u_i\) and \(u_j\).

I’ll also ignore the possibility of trade before existence, or Rawlsian veils of ignorance. If you are a \(u_i\) maximiser, but you could have been a \(u_j\) maximiser if things had been different, then you have no responsibility to increase \(u_j\). Similarly, if there are \(u_j\) maximisers out there, then you have no responsibility to maximiser \(u_j\) without getting any \(u_i\) increases out of that. See this post for more on that.

Changing that last assumption could radically alter the nature of acausal trade - potentially reducing it to simply maximising a universal prior utility function. See this post for more on that behaviour.



NEW LINKS

NEW POSTS

NEW DISCUSSION POSTS

RECENT COMMENTS

[Note: This comment is three
by Ryan Carey on A brief note on factoring out certain variables | 0 likes

There should be a chat icon
by Alex Mennen on Meta: IAFF vs LessWrong | 0 likes

Apparently "You must be
by Jessica Taylor on Meta: IAFF vs LessWrong | 1 like

There is a replacement for
by Alex Mennen on Meta: IAFF vs LessWrong | 1 like

Regarding the physical
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

I think that we should expect
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

I think I understand your
by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

This seems like a hack. The
by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

After thinking some more,
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Yes, I think that we're
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

My intuition is that it must
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

To first approximation, a
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Actually, I *am* including
by Vadim Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Yeah, when I went back and
by Alex Appel on Optimal and Causal Counterfactual Worlds | 0 likes

> Well, we could give up on
by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

RSS

Privacy & Terms