Is there any rigorous work on using anthropic uncertainty to prevent situational awareness / deception? — LessWrong