x
Hoodwinked: Evaluating Deception Capabilities in Large Language Models — LessWrong