TLDR: CAIS is distributing $250,000 in prizes for benchmarks that empirically assess AI safety. This project is supported by Schmidt Sciences, submissions are open until February 25th, 2025. Winners will be announced April 25th, 2025. To view additional info about the competition, including submission guidelines and FAQs, visit https://www.mlsafety.org/safebench If...
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Subscribe here to receive future versions. --- Cybersecurity Challenges in AI Safety Meta accidentally leaks a language model to the public. Meta’s newest language model, LLaMa,...
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Subscribe here to receive future versions. --- Policy Proposals for AI Safety Critical industries rely on the government to protect consumer safety. The FAA approves new...
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Subscribe here to receive future versions. --- ChaosGPT and the Rise of Language Agents Chatbots like ChatGPT usually only respond to one prompt at a time,...
The Center for AI Safety just launched its first AI Safety Newsletter. The newsletter is designed to inform readers about developments in AI and AI safety. No technical background required. Subscribe here to receive future versions. First edition below: --- Growing concerns about rapid AI progress Recent advancements in AI...
TLDR We're announcing a new course designed to introduce students with a background in machine learning to the most relevant concepts in empirical ML-based AI safety. The course is available publicly here. Background AI safety is a small but rapidly growing field, and both younger and more experienced researchers are...
TLDR We are announcing a $20k bounty for publicly-understandable explainers of AI safety concepts. We are also releasing the results of the AI Safety Arguments competition. Background Of the technologists, ML researchers, and policymakers thinking about AI, very few are seriously thinking about AI existential safety. This results in less...