Also, a new open letter advocating for pro-human values and control over AI development Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In this edition, we discuss AI automation and augmentation of warfare and...
Also, Anthropic Removes a Core Safety Commitment Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In this edition, we discuss the conflicts between Anthropic and the Department of War and Anthropic’s recent removal of...
Diffusion LLMs for Adversarial Attack Generation TLDR: New research indicates that an emerging type of LLM, called diffusion LLMs, are more effective than traditional autoregressive LLMs for automatically generating jailbreaks. Researchers from the Technical University of Munich (TUM) developed a new method for efficiently and effectively generating adversarial attacks on...
Measuring General AI Abilities TLDR: New metrics say that frontier AIs get 57% on tests for general intelligence and are able to do 2.5% of remote freelance-type work. Many benchmarks measure AIs on useful knowledge and abilities, but there is a measurement gap around some of the most important capability...
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In this edition: A new benchmark measures AI automation; 50,000 people, including top AI scientists, sign an open letter calling for a superintelligence moratorium. Listen to...
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In this edition: A new bill in the Senate would hold AI companies liable for harms their products create; China tightens its export controls on rare...
r AI Safety. We discuss developments in AI and AI safety. No technical background required. In this edition: California’s legislature sent SB-53—the ‘Transparency in Frontier Artificial Intelligence Act’—to Governor Newsom’s desk. If signed into law, California would become the first US state to regulate catastrophic risk. Listen to the AI...