On the eve of superintelligence, real AI safety is a nonexistent field. The AI companies have embraced something else: safety through psychoanalysis (for lack of a better term). Their safety team concocts various test scenarios for their AI in order to learn various traits of the AI’s “personality”. They even...
My beef with Anthropic I've long felt that while Anthropic is the most safety-conscious of the frontier AI companies, they're also the most hypocritical enablers of the whole reckless enterprise. By framing themselves as the "good sport" in the race, the one who's encouraging everyone else to "race them to...
Is focusing on corrigibility our best shot at getting to ASI alignment? Max Harms and Jeremy Gillen are current and former MIRI alignment researchers who both see superintelligent AI as an imminent extinction threat, but disagree about Max's proposal of Corrigibility as Singular Target (CAST). Max thinks focusing on corrigibility...
Mutual-Knowledgeposting The purpose of this post is to build mutual knowledge that many (most?) of us on LessWrong support If Anyone Builds It, Everyone Dies. Inside of LW, not every user is a long-timer who's already seen consistent signals of support for these kinds of claims. A post like this...
My interview with Eliezer Yudkowsky for If Anyone Builds It, Everyone Dies launch week is out! Video Timestamps * 00:00:00 — Eliezer Yudkowsky Intro * 00:01:25 — Recent validation of Eliezer's ideas * 00:03:46 — Sh*t now getting real * 00:08:47 — Eliezer’s rationality teachings * 00:10:39 — Rationality Lesson...
Dr. @Steven Byrnes is one of the few people who both understands why alignment is hard, and is taking a serious technical shot at solving it. He's the author of these recently popular posts: * Foom & Doom 1: "Brain in a box in a basement" * Foom & Doom...
Carl Feynman (@Carl Feynman) is a career-long AI Engineer, M.S. in Computer Science from MIT, and son of Richard Feynman. He’s a lifelong rationalist, has known Eliezer Yudkowsky since the ‘90s, and he witnessed Eliezer’s AI doom argument taking shape before most of us were paying any attention. I interviewed...