There have been multiple recent calls for the automation of AI safety and alignment research. There are likely many people who would like to contribute to this space, but would benefit from clear directions for how to do so. Stemming from a recent SPAR project and in light of limitations...
In light of the recent news from Mechanize/Epoch and the community discussion it sparked, I'd like to open a conversation about a question some of us grapple with: What constitutes a net-positive AI startup from an AI safety perspective, and what steps can founders take to demonstrate goodwill and navigate...
This post is broken down into two parts: 1. Which AI productivity tools am I currently using (as an alignment researcher)? 2. Why does it currently feel hard to spend +$1000/month on AI to increase one's productivity drastically? Which AI productivity tools am I currently using? Let's get right to...
I've been going through the FAR AI videos from the alignment workshop in December 2023. I'd like people to discuss their thoughts on Shane Legg's 'necessary properties' that every AGI safety plan needs to satisfy. The talk is only 5 minutes, give it a listen: Otherwise, here are some of...
Apply to work on this project with me at AI Safety Camp 2024 before 1st December 2023. Summary Future prosaic AIs will likely shape their own development or that of successor AIs. We're trying to make sure they don't go insane. There are two main ways AIs can get better:...
[This post summarizes some of the work done by Owen Dudney, Roman Engeler and myself (Quintin Pope) as part of the SERI MATS shard theory stream.] TL;DR Future prosaic AIs will likely shape their own development or that of successor AIs. We're trying to make sure they don't go insane....
New article in Time Ideas by Eliezer Yudkowsky. Here’s some selected quotes. In reference to the letter that just came out (discussion here): > We are not going to bridge that gap in six months. > > It took more than 60 years between when the notion of Artificial Intelligence...