by Ryan Carey 340 days ago | Jessica Taylor likes this | link | parent This result features in the paper by Piccione and Rubeinstein that introduced the absent-minded driver problem [1]. Philosophers like decision theories that self-ratify, and this is indeed a powerful self-ratification principle. This self-ratification principle does however rely on SIA probabilities assuming the current policy. We have shown that conditioning on your current policy, you will want to continue on with your current policy. i.e. the policy will be a Nash Equilibrium. There can be Nash Equilibria for other policies $$\pi'$$ however. The UDT policy will by definition equal or beat these from the ex ante point of view. However, others can achieve higher expected utility conditioning on the initial observation i.e. higher $$SIA_{\pi'}(s|o)Q_{\pi'}(s,a)$$. This apparent paradox is discussed in [2] [3], and seems to reduce to disagreement over counterfactual mugging. So why do we like the UDT solution over solutions that are more optimal locally, and that also locally self-ratify? Obviously we want to avoid resorting so circular reasoning (i.e. it gets the best utility ex ante). I think there are some okay reasons: it is reflectively stable (i.e. will not self-modify, will not hide future evidence) and it makes sense assuming modal realism or many worlds interpretation (then we deem it parochial to focus on any reference frame other than equal weighting across the whole wavefunction/universe) it makes sense if we assume that self-location somehow does not it’s simpler (utility function given weighting 1 across all worlds). In principle, UDT can also include the locally optimal it transfers better to scenarios without randomization as in Nate + Ben Levenstein’s forthcoming [4]. I imagine there are more good arguments that I don’t yet know. p19 Piccione, Michele, and Ariel Rubinstein. “On the interpretation of decision problems with imperfect recall.” Games and Economic Behavior 20.1 (1997): 3-24. Schwarz, Wolfgang. “Lost memories and useless coins: revisiting the absentminded driver.” Synthese 192.9 (2015): 3011-3036. http://lesswrong.com/lw/3dy/has_anyone_solved_psykoshs_nonanthropic_problem/ Cheating Death in Damascus / Nate Soares and Ben Levenstein / Forthcoming

### NEW DISCUSSION POSTS

This is exactly the sort of
 by Stuart Armstrong on Being legible to other agents by committing to usi... | 0 likes

When considering an embedder
 by Jack Gallagher on Where does ADT Go Wrong? | 0 likes

The differences between this
 by Abram Demski on Policy Selection Solves Most Problems | 0 likes

Looking "at the very
 by Abram Demski on Policy Selection Solves Most Problems | 0 likes

 by Paul Christiano on Policy Selection Solves Most Problems | 1 like

>policy selection converges
 by Stuart Armstrong on Policy Selection Solves Most Problems | 0 likes

Indeed there is some kind of
 by Vadim Kosoy on Catastrophe Mitigation Using DRL | 0 likes

Very nice. I wonder whether
 by Vadim Kosoy on Hyperreal Brouwer | 0 likes

Freezing the reward seems
 by Vadim Kosoy on Resolving human inconsistency in a simple model | 0 likes

Unfortunately, it's not just
 by Vadim Kosoy on Catastrophe Mitigation Using DRL | 0 likes

>We can solve the problem in
 by Wei Dai on The Happy Dance Problem | 1 like

Maybe it's just my browser,
 by Gordon Worley III on Catastrophe Mitigation Using DRL | 2 likes

At present, I think the main
 by Abram Demski on Looking for Recommendations RE UDT vs. bounded com... | 0 likes

In the first round I'm
 by Paul Christiano on Funding opportunity for AI alignment research | 0 likes

Fine with it being shared
 by Paul Christiano on Funding opportunity for AI alignment research | 0 likes