by Patrick LaVictoire 883 days ago | link | parent I can prove the property that for each hypothesis $$A()=a$$ there is at most one $$u$$ such that $$U()=u$$ has a high valuation (for sufficiently high PA+N), with the following caveat: it can sometimes take many steps to prove that $$u\neq u'$$ in PA+N, so we’ll need to include the length of that proof in our bound. In what follows, we will take all subscripts of $$d$$ and $$\nu$$ to be $$PA+N, A()=a$$ for $$N$$ large. For any $$\phi$$, $$d(\bot) - d(\neg\phi)\leq d(\phi)\leq d(\bot)$$ and thus $1 - \frac{d(\phi)}{d(\bot)} \leq \nu(\phi) \leq \frac{d(\bot)}{d(\phi)+d(\bot)}.$ Also, $$d(U()=u)+d(U()=u')+d(u\neq u')\geq d(\bot)$$. This implies $$\max\{d(U()=u),d(U()=u')\}\geq \frac12(d(\bot)-d(u\neq u))$$, which implies $\min\{\nu(U()=u),\nu(U()=u')\}\leq \min\{\frac{d(\bot)}{d(U()=u)+d(\bot)},\frac{d(\bot)}{d(U()=u')+d(\bot)}\} \leq \frac{2d(\bot)}{3d(\bot)-d(u\neq u')}.$ So we see that $$\nu(U()=u)$$ and $$\nu(U()=u')$$ cannot both be significantly larger than 2/3 if there is a short proof that $$u\neq u'$$.

### NEW DISCUSSION POSTS

[Delegative Reinforcement
 by Vadim Kosoy on Stable Pointers to Value II: Environmental Goals | 1 like

Intermediate update: The
 by Alex Appel on Further Progress on a Bayesian Version of Logical ... | 0 likes

Since Briggs [1] shows that
 by 258 on In memoryless Cartesian environments, every UDT po... | 2 likes

This doesn't quite work. The
 by Nisan Stiennon on Logical counterfactuals and differential privacy | 0 likes

I at first didn't understand
 by Sam Eisenstat on An Untrollable Mathematician | 1 like

This is somewhat related to
 by Vadim Kosoy on The set of Logical Inductors is not Convex | 0 likes

This uses logical inductors
 by Abram Demski on The set of Logical Inductors is not Convex | 0 likes

Nice writeup. Is one-boxing
 by Tom Everitt on Smoking Lesion Steelman II | 0 likes

Hi Alex! The definition of
 by Vadim Kosoy on Delegative Inverse Reinforcement Learning | 0 likes

A summary that might be
 by Alex Appel on Delegative Inverse Reinforcement Learning | 1 like

I don't believe that
 by Alex Appel on Delegative Inverse Reinforcement Learning | 0 likes

This is exactly the sort of
 by Stuart Armstrong on Being legible to other agents by committing to usi... | 0 likes

When considering an embedder
 by Jack Gallagher on Where does ADT Go Wrong? | 0 likes

The differences between this
 by Abram Demski on Policy Selection Solves Most Problems | 1 like

Looking "at the very
 by Abram Demski on Policy Selection Solves Most Problems | 0 likes