LESSWRONG
LW

1145
Ryan Kidd
2567Ω72222202
Message
Dialogue
Subscribe
  • Co-Executive Director at ML Alignment & Theory Scholars Program (2022-present)
  • Co-Founder & Board Member at London Initiative for Safe AI (2023-present)
  • Manifund Regrantor (2023-present)  |  RFPs here
  • Advisor, Catalyze Impact (2023-present)  |  ToC here
  • Advisor, AI Safety ANZ (2024-present)
  • Ph.D. in Physics at the University of Queensland (2017-2023)
  • Group organizer at Effective Altruism UQ (2018-2021)

Give me feedback! :)

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
3Ryan Kidd's Shortform
3y
123
Ryan Kidd's Shortform
Ryan Kidd3mo*1270

80% of MATS alumni who completed the program before 2025 are still working on AI safety today, based on a survey of all available alumni LinkedIns or personal websites (242/292 ~ 83%). 10% are working on AI capabilities, but only ~6 at a frontier AI company (2 at Anthropic, 2 at Google DeepMind, 1 at Mistral AI, 1 extrapolated). 2% are still studying, but not in a research degree focused on AI safety. The last 8% are doing miscellaneous things, including non-AI safety/capabilities software engineering, teaching, data science, consulting, and quantitative trading.

Of the 193+ MATS alumni working on AI safety (extrapolated: 234):

  • 34% are working at a non-profit org (Apollo, Redwood, MATS, EleutherAI, FAR.AI, MIRI, ARC, Timaeus, LawZero, RAND, METR, etc.);
  • 27% are working at a for-profit org (Anthropic, Google DeepMind, OpenAI, Goodfire, Meta, etc.);
  • 18% are working as independent researchers, probably with grant funding from Open Philanthropy, LTFF, etc.;
  • 15% are working as academic researchers, including PhDs/Postdocs at Oxford, Cambridge, MIT, ETH Zurich, UC Berkeley, etc.;
  • 6% are working in government agencies, including in the US, UK, EU, and Singapore.

10% of MATS alumni co-founded an active AI safety start-up or team during or after the program, including Apollo Research, Timaeus, Simplex, ARENA, etc.

Errata: I mistakenly included UK AISI in the "non-profit AI safety organization" category instead of "government agency". I also mistakenly said that the ~6 alumni working on AI capabilities at frontier AI companies were all working on pre-training.

Reply132
Ryan Kidd's Shortform
Ryan Kidd1y*291

Why does the AI safety community need help founding projects?

  1. AI safety should scale
    1. Labs need external auditors for the AI control plan to work
    2. We should pursue many research bets in case superalignment/control fails
    3. Talent leaves MATS/ARENA and sometimes struggles to find meaningful work for mundane reasons, not for lack of talent or ideas
    4. Some emerging research agendas don’t have a home
    5. There are diminishing returns at scale for current AI safety teams; sometimes founding new projects is better than joining an existing team
    6. Scaling lab alignment teams are bottlenecked by management capacity, so their talent cut-off is above the level required to do “useful AIS work”
  2. Research organizations (inc. nonprofits) are often more effective than independent researchers
    1. “Block funding model” is more efficient, as researchers can spend more time researching, rather than seeking grants, managing, or other traditional PI duties that can be outsourced
    2. Open source/collective projects often need a central rallying point (e.g., EleutherAI, dev interp at Timaeus, selection theorems and cyborgism agendas seem too delocalized, etc.)
  3. There is (imminently) a market for for-profit AI safety companies and value-aligned people should capture this free energy or let worse alternatives flourish
    1. If labs or API users are made legally liable for their products, they will seek out external red-teaming/auditing consultants to prove they “made a reasonable attempt” to mitigate harms
    2. If government regulations require labs to seek external auditing, there will be a market for many types of companies
    3. “Ethical AI” companies might seek out interpretability or bias/fairness consultants
  4. New AI safety organizations struggle to get funding and co-founders despite having good ideas
    1. AIS researchers are usually not experienced entrepeneurs (e.g., don’t know how to write grant proposals for EA funders, pitch decks for VCs, manage/hire new team members, etc.)
    2. There are not many competent start-up founders in the EA/AIS community and when they join, they don’t know what is most impactful to help
    3. Creating a centralized resource for entrepeneurial education/consulting and co-founder pairing would solve these problems
Reply
Ryan Kidd's Shortform
Ryan Kidd1y*482

I am a Manifund Regrantor. In addition to general grantmaking, I have requests for proposals in the following areas:

  • Funding for AI safety PhDs (e.g., with these supervisors), particularly in exploratory research connecting AI safety theory with empirical ML research.
  • An AI safety PhD advisory service that helps prospective PhD students choose a supervisor and topic (similar to Effective Thesis, but specialized for AI safety).
  • Initiatives to critically examine current AI safety macrostrategy (e.g., as articulated by Holden Karnofsky) like the Open Philanthropy AI Worldviews Contest and Future Fund Worldview Prize.
  • Initiatives to identify and develop "Connectors" outside of academia (e.g., a reboot of the Refine program, well-scoped contests, long-term mentoring and peer-support programs).
  • Physical community spaces for AI safety in AI hubs outside of the SF Bay Area or London (e.g., Japan, France, Bangalore).
  • Start-up incubators for projects, including evals/red-teaming/interp companies, that aim to benefit AI safety, like Catalyze Impact, Future of Life Foundation, and YCombinator's request for Explainable AI start-ups.
  • Initiatives to develop and publish expert consensus on AI safety macrostrategy cruxes, such as the Existential Persuasion Tournament and 2023 Expert Survey on Progress in AI (e.g., via the Delphi method, interviews, surveys, etc.).
  • Ethics/prioritization research into:
    • What values to instill in artificial superintelligence?
    • How should AI-generated wealth be distributed?
    • What should people do in a post-labor society?
    • What level of surveillance/restriction is justified by the Unilateralist's Curse?
    • What moral personhood will digital minds have?
    • How should nations share decision making power regarding transformative AI?
  • New nonprofit startups that aim to benefit AI safety.
Reply11
AI safety undervalues founders
Ryan Kidd8h20

Thanks for sharing, Akshyae! Based on the DMs I received after posting this, I think your experience is unfortunately common. Great job sticking at it and launching Explainable!

Reply
AI safety undervalues founders
Ryan Kidd10h20

Can you share which?

Reply
AI safety undervalues founders
Ryan Kidd10h40

True, but we accepted 75% of all scholars into the 6-month extension last program, so the pressure might not be that large now.

Reply
AI safety undervalues founders
Ryan Kidd2d30

Of note: when I first approached you about becoming a MATS mentor, I don't think you had significant field-building or mentorship experience and had relatively few papers. Since then, you have become one of the most impactful field-builders, mentors, and researchers in AI safety, by my estimation! This is a bet I would take again.

Reply11
AI safety undervalues founders
Ryan Kidd2d90

Edit: I mistakenly said "27% at frontier labs" when I should have said "27% at for-profit companies". Also, note that this is 27% of those working on AI safety (80%), so 22% of all alumni.

Reply1
AI safety undervalues founders
Ryan Kidd2d10

In regards to adversarial selection, we can compare MATS to SPAR. SPAR accepted ~300 applicants in their latest batch, ~3x MATS (it's easier to scale if you're remote, don't offer stipends, and allow part-timers). I would bet that the average research impact of SPAR participants is significantly lower than that of MATS, though there might be plenty of confounders here. It might be worth doing a longitudinal study here comparing various training programs' outcomes over time, including PIBBSS, ERA, etc.

I think your read of the situation re. mentor ratings is basically correct: increasingly many MATS mentors primarily care about research execution ability (generally ML), not AI safety strategy knowledge. I see this as a feature, not a bug, but I understand why you disagree. I think you are prioritizing a different skillset than most mentors that our mentor selection committee rates highly. Interestingly, most of the technical mentors that you rate highly seem to primarily care about object-level research ability and think that strategy/research taste can be learned on the job!

Note that I think the pendulum might start to swing back towards mentors valuing high-level AI safety strategy knowledge as the Iterator archetype is increasingly replaced/supplemented by AI. The Amplifier archetype seems increasingly in-demand as orgs scale, and we might see a surge in Connectors as AI agents improve to the point that their theoretical ideas are more testable. Also note that we might have different opinions on the optimal ratio of "visionaries" vs. "experimenters" in an emerging research field.

Reply
AI safety undervalues founders
Ryan Kidd2d30

I like this comment. I think it's easy to overfit on the most salient research agendas, especially if there are echo chambers and tight coupling between highly paid frontier AI staff and nonprofit funders. The best way I know to combat this at MATS is:

  • Maintain a broad church of AI safety research, including deliberately making mentor "diversity picks" and choosing a mentor selection committee that contains divergent thinkers. As another example, I think Constellation has done a good job recently at expanding member diversity and reducing echo chambers.
  • Requiring that COIs be declared and mitigated, including along reporting chains, at the same organization, with romantic/sexual partners, and with frequent research collaborators.
  • Encouraging "scout mindset" and "reasoning transparency", especially among people with divergent beliefs. I think this is a large strength of MATS: we are a melting pot for ideas and biases.

Note that I expect overfitting to decrease with further scale and diversity, given the above practices are adhered to!

Reply1
Load More
96AI safety undervalues founders
2d
66
47Apply to MATS 9.0!
2mo
0
22MATS 8.0 Research Projects
Ω
2mo
Ω
0
8MATS is hiring!
7mo
0
64Apply to MATS 8.0!
8mo
5
27MATS Spring 2024 Extension Retrospective
9mo
1
18[Job ad] LISA CEO
9mo
4
96Implications of the inference scaling paradigm for AI safety
10mo
70
44MATS mentor selection
10mo
12
10[Job Ad] MATS is hiring!
1y
0
Load More
MATS Program
2 years ago
(+255/-37)
MATS Program
2 years ago
(+14/-46)