Crossposted from the AI Alignment Forum. May contain more technical jargon than usual.
This is a linkpost for http://mathb.in/39392

New to LessWrong?

New Comment
1 comment, sorted by Click to highlight new comments since: Today at 4:01 AM

This might combine well with the setup I thought of for studying bounded rationality / approximate utility maximizers. Namely, the setup is like that for UDT, but your algorithm has to pay 1 utilon for every call to its halting oracle. What algorithm might you use for this game?

(I find it easier to imagine that the algorithm is proving things only about the utility of the action not counting the computational costs.)