Retrospective: 12 [sic] Months Since MIRI
it's now been 15 months since MIRI but I just remembered that three separate people have told me they liked this post despite my not cross-posting it, so I am now cross-posting it. Not written with the intention of being useful to any particular audience, just collecting my thoughts on this past year's work. September-December 2023: Orienting and Threat Modeling Until September, I was contracting full time for a project at MIRI. When the MIRI project ended, I felt very confused about lots of things in AI safety. I didn't know what sort of research would be useful for making AI safe, and I also didn't really know what my cruxes were for resolving that confusion. When people asked what I was thinking about, I usually responded with some variation of the following: > MIRI traditionally says that even though their research is hard and terrible, everyone else's research is not at all on track to have a shot at making safe AI, and so MIRI research is still the best option available. > > After working with MIRI for a year, I certainly agree that their research is hard and terrible. I have yet to be convinced that everyone else's research is worse. I'm working on figuring that out. So I set out to investigate why MIRI thinks their research is the only kind of thing that has a shot, and if they're right or wrong about that. My plan? Read a bunch of stuff, think about it, write down my thoughts, talk to lots of people about it, try to get a foothold on the vague mass of confusion in my head. To my credit, I did anticipate that this could eat up infinite time and effort if I let it, so I decided to call a hard "stop and re-evaluate" at the end of September. To my discredit, when the end-of-month deadline came I took a moment to reflect on what I'd learned, and I realized I still didn't know what to do next. I thought that all I needed was a little more thinking bro and all the pieces would snap into place. Just a little more thinking bro. No need to set a next stop-and-r