Question/​Issue with the 5/​10 Problem

https://www.lesswrong.com/posts/aAMLFc9AogbJwN8xZ/question-issue-with-the-5-10-problem

I’m not sure if the 5⁄10 problem and the surrounding Löbian uncertainty is still an issue/​area of research, but I’ve been struggling with the validity of this argument lately—I doubt this is a novel point so if this is addressed somewhere I’d appreciate being corrected. On the off chance that this hasn’t been explained elsewhere I’d be really interested to hear peoples’ thoughts The proof as I understand it is roughly as below, with "A" referring to the agent’s output, "U" referring to the environment’s output, and "□" standing for the provability predicate:

Comment

https://www.lesswrong.com/posts/aAMLFc9AogbJwN8xZ/question-issue-with-the-5-10-problem?commentId=RJg9GnFaNfg223yp9

The core of the 5-and-10 problem is not specific to a particular formalization or agent algorithm. It’s fundametally the question of what’s going on with agent’s reasoning inside the 5 world. In the 10 world, agent’s reasoning proceeds in a standard way, perhaps the agent considers both the 5 and 10 worlds, evaluates them, and decides to go with 10. But what might the agent be thinking in the 5 world, so that it ends up making that decision? And if the agent in the 10 world is considering the 5 world, what does the agent in the 10 world think about the thinking of the agent in the 5 world, and about what that implies in general?

How this happens is a test for decision making algorithms, as it might lead to a breakdown along the lines of the 5-and-10 problem, or to a breakdown of an informal model of how a particular algorithm works. The breakdown is not at all inevitable, and usually the test can’t even be performed without changing the algorithm to make it possible, in which case we’ve intentionally broken the algorithm in an interesting way that might tell us something instructive.

In the post, what agent algorithm are you testing? Note that agent’s actions are not the same thing as agent’s knowledge of them. Proving A = 5 in a possibly inconsistent system is not the same thing as actually doing 5 (perhaps the algorithm explicitly says to do 10 upon proving A = 5, which is the chicken rule; there is no relevant typo in this parenthetical).

Comment

https://www.lesswrong.com/posts/aAMLFc9AogbJwN8xZ/question-issue-with-the-5-10-problem?commentId=Cof9aiw3s6z8tbCAp

Yeah sure, like there’s a logical counterfactual strand of the argument but that’s not the topic I’m really addressing here—I find those a lot less convincing so my issue here is around the use of Lobian uncertainty specifically. There’s an step very specific to this species of argument that proving that □P will make P true when P is about the outcomes of the bets, because you will act based on the proof of P. This is invoking Lob’s Theorem in a manner which is very different from the standard counterpossible principle of explosion stuff. And I’m really wanting to discuss that step specifically because I don’t think it’s valid, and if the above argument is still representative of at least a strand of relevant argument then I’d be grateful for some clarification on how (3.) is supposed to be provable by the agent, or how my subsequent points are invalid.

https://www.lesswrong.com/posts/aAMLFc9AogbJwN8xZ/question-issue-with-the-5-10-problem?commentId=DJkjBFpMvdYQuXozJ

From the link

"I have to decide between $5 and $10. Suppose I decide to choose $5. I know that I’m a money-optimizer, so if I do this, $5 must be more money than $10, so this alternative is better. Therefore, I should choose $5." To me there is a sleight of hand there. The statement "I know that I’m a money-optimizer" is not a mathematical statement, but an empirical one, it can be tested through one’s actions. If you take $5 instead of $10, you are not a money-optimizer, even if you initially think you are, and that’s something, as an agent, you can learn about yourself by observing your actions.