Have you considered doing e.g. 80k careers coaching or otherwise sitting down with a fieldbuilder like Ryan Kidd or James Fox or a funder like OpenPhil or Jaan Tallinn? Seems like you're looking for someone with a global perspective who knows what the field needs so you can shape yourself to that mould and those are the people that would know
Thank you for the suggestion. This is something I've both considered and done in practice, having received coaching from some of the people/orgs you've mentioned. I should get in touch with OP to see if they'd like to fund any of this work though, as that could be a useful feedback signal.
Thank you for sharing your knowledge on existing work in the below comment as well - it's super helpful!
Re: 10 I think the dual-use of capabilities is widely recognized by influential persons e.g. David Krueger, Dan Hendrycks, Nitarshan of UK AISI, etc. The overepresentation of evals in the global portfolio is mostly because it's tractable, easy to report on, doesn't require a degree in CS, low-resource (you don't need your own inference hardware or access to model weights to do evals) and seems useful. CF mechinterp: also known to be dual use but still very tractable and easy to measure and report on
Hello everyone! 👋
At the start of year, I was trying to figure out what my next goals should be regarding AI safety. I ended up making a list of 130 small, concrete ideas ranging all kinds of domains. A lot of them are about gathering and summarizing information, skilling up myself, and helping others be more effective.
I needed a rough way to prioritize them, so I built a prioritization framework inspired by 80,000 Hours. I won't go into the details here, but basically it allowed me to filter out the top 12 ideas - one for each month.
In the first half of the year I did the most urgent ones, and now I'm in Q2 of the Eisenhower matrix. The shortlist still has 12 things to do, since I've come up with new ideas during the first half of 2025. However, I need a bit more input to narrow down the next steps.
This is where you come in. I'd like to hear your thoughts on the shortlist and the relative importance of its tasks. Especially valuable would be to know if something is already being done by someone who can get it done - then I can skip it and get closer to the margin.
1. Visualize the distribution of people in research/governance fields; make the safety "portfolio" clearer and more tangible
2. Find out what % of Schmidt Futures or UK AISI's safety-earmarked budget actually gets used
3. Catalogue and evaluate practical plans to deter or influence RSI
4. Investigate and post about how crucial it is to get mainstream media out of the competition/race framing
5. Come up with a response to a friend's argument over lack of agency
6. Talk about evaluations at Chiba University's AI safety workshop
7. Offer productivity tools and coaching to AI safety researchers
8. Do a calculation on whether it's net positive for Reaktor to work on AI safety
9. Write a LW/AF post convincing people in safety to take lower salaries
10. Write about how evals might actually push capabilities, as model developers take on new challenges
11. Write an article on how easy to shut down AGI/ASI systems should be
12. Write a LW/AF post about why we should divest half of interpretability
If you have thoughts or context to share about any of these, that would be much appreciated. Thank you!
Doing the prioritization was a useful exercise in itself: I noticed most of my ideas seemed ok at the time of having them, but were actually pretty bad upon further inspection. Being able to prioritize and narrow them down by 90% was valuable. It even helped me generate better ideas afterwards, since I can easily compare new ideas to the existing ones.
This also resolved another problem of mine: ideas and advice from other people, while useful, often fails to account for my personal situation (strong desire to live in Japan) and unique opportunities (broad network across cultural and language borders).
Putting numbers to all the ideas made it easier to understand and articulate what I value and am motivated by. It's also nice to have immediate feedback about an idea in an unambiguous form, and I like being able to forget about all the random details: I can simply add notes to my spreadsheet and refer to them when needed.
However, now I've finished the highest-priority tasks and need to figure out what to do in the latter half of the year. My aim has been to do one thing per month; that seems sustainable with my other commitments. I'm hoping that sharing this publicly will encourage others to seek more feedback about their priorities as well.