Safe ASI Is Achievable: The Finite Game Argument
A few days ago, Anthropic dropped the central pledge of its Responsible Scaling Policy, the promise that it would never train an AI system unless it could guarantee in advance that its safety measures were adequate. The stated reason: unilateral safety commitments don't make sense when competitors are racing ahead...