Another day, another METR graph update. METR said on X: > We estimate that Claude Opus 4.6 has a 50%-time-horizon of around 14.5 hours (95% CI of 6 hrs to 98 hrs) on software tasks. While this is the highest point estimate we’ve reported, this measurement is extremely noisy because...
People's Clawdbots now have their own AI-only Reddit-like Social Media called Moltbook and they went from 1 agent to 36k+ agents in 72 hours. As Karpathy puts it: > What's currently going on at @moltbook is genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently. People's Clawdbots (moltbots,...
I just released a 22-minute long documentary about the hunger strike to stop the AI race, a protest that was featured in publications such as The Verge, The Telegraph and even SkyNews, and received internal support from employees at DeepMind. A few months ago, during the protest, I wrote a...
Andrej Karpathy posted 12 hours ago (emphasis mine): > I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between. I have a sense that I could be 10X more powerful if I just...
An updated METR graph including Claude Opus 4.5 was just published 3 hours ago on X by METR (source): Same graph but without the log (source): Thread from METR on X (source): > We estimate that, on our tasks, Claude Opus 4.5 has a 50%-time horizon of around 4 hrs...
Are there any public cost-effectiveness analyses of different AI Safety charities? For instance, I'm aware of Larks' AI Alignment Literature Review and Charity Comparison, but he didn't include any concrete impact measures and stopped doing it in 2021. I'm looking for things like "donating $10M to this org would reduce...
Adapted from a Manifund proposal I announced yesterday. In the past two weeks, I have been posting daily AI-Safety-related clips on TikTok and YouTube reaching more than 1M people. Screenshot from my TikTok channel taken a couple days ago I'm doing this because I believe short-form AI Safety content is...