Note: This is an automated crosspost from METR. The bot selects content from many AI safety-relevant sources. Not affiliated with the authors or their organization.
Many frontier AI safety policies from scaling labs (e.g. OpenAI’s Preparedness Framework, Google DeepMind’s Frontier Safety Framework, etc.), as well as past work by third party evaluators including UK AISI, Apollo Research, and METR, focus on pre-deployment testing – ensuring that the AI model is safe and that the lab has sufficient security before the lab deploys the model to the public.
Such pre-deployment safety evaluations are standard for a wide variety of products across many industries, where the primary risk of the product is to the consumer (see,... (read 806 more words →)