by Jack Gallagher 592 days ago | link | parent | on: Where does ADT Go Wrong? When considering an embedder $$F$$, in universe $$U$$, in response to which SADT picks policy $$\pi$$, I would be tempted to apply the following coherence condition: $E[F(\pi)] = E[F(DDT)] = E[U]$ (all approximately of course) I’m not sure if this would work though. This is definitely a necessary condition for reasonable counterfactuals, but not obviously sufficient. A potentially useful augmentation is to use absolute expected difference: $E[|F(\pi) - F(DDT)|] = E[|F(DDT) - U|] = 0$ reply
 by Jack Gallagher 1097 days ago | link | parent | on: A new proposal for logical counterfactuals By censoring I mean a specific technique for forcing the consistency of a possibly inconsistent set of axioms. Suppose you have a set of deduction rules $$D$$ over a language $$\ell$$. You can construct a function $$f_D : P(\ell) \to P(\ell)$$ that takes a set of sentences $$S$$ and outputs all the sentences that can be proved in one step using $$D$$ and the sentences in $$S$$. You can also construct a censored $$f'_D$$ by letting $$f'_D(S) = \{\phi\ |\ \phi \in f_D(S) \wedge \neg \phi \not\in S\}$$. reply

### NEW DISCUSSION POSTS

[Note: This comment is three
 by Ryan Carey on A brief note on factoring out certain variables | 0 likes

There should be a chat icon
 by Alex Mennen on Meta: IAFF vs LessWrong | 0 likes

Apparently "You must be
 by Jessica Taylor on Meta: IAFF vs LessWrong | 1 like

There is a replacement for
 by Alex Mennen on Meta: IAFF vs LessWrong | 1 like

Regarding the physical
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

I think that we should expect
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

I think I understand your
 by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

This seems like a hack. The
 by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

After thinking some more,
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Yes, I think that we're
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

My intuition is that it must
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

To first approximation, a
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Actually, I *am* including
 by Vanessa Kosoy on The Learning-Theoretic AI Alignment Research Agend... | 0 likes

Yeah, when I went back and
 by Alex Appel on Optimal and Causal Counterfactual Worlds | 0 likes

> Well, we could give up on
 by Jessica Taylor on The Learning-Theoretic AI Alignment Research Agend... | 0 likes