Evaluating strategic reasoning in GPT models — LessWrong