LESSWRONG
LW

Benjamin Hilton
400Ω69610
Message
Dialogue
Subscribe

Head of Alignment at UK AI Security Institute (AISI). Previously 80,000 Hours, HM Treasury, Cabinet Office, Department for International Trade, Imperial College London.

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
The Alignment Project Research Agenda
UK AISI Alignment Team: Debate Sequence
Why I think it's net harmful to do technical safety research at AGI labs
Benjamin Hilton2y3-2

[x-posted from EA forum]

Hi Remmelt,

Thanks for sharing your concerns, both with us privately and here on the forum. These are tricky issues and we expect people to disagree about how to about how to weigh all the considerations — so it’s really good to have open conversations about them.

Ultimately, we disagree with you that it's net harmful to do technical safety research at AGI labs. In fact, we think it can be the best career step for some of our readers to work in labs, even in non-safety roles. That’s the core reason why we list these roles on our job board.

We argue for this position extensively in my article on the topic (and we only list roles consistent with the considerations in that article).

Some other things we’ve published on this topic in the last year or so:

  • A range of opinions from anonymous experts about the upsides and downsides of working on AI capabilities
  • How policy roles in AI companies can be valuable for career capital and for direct impact (as well as the potential downsides)
  • We recently released a podcast episode with Nathan Labenz on some of the controversy around OpenAI, including his concerns about some of their past safety practices, whether ChatGPT’s release was good or bad, and why its mission of developing AGI may be too risky.

Benjamin

Reply
12Research Areas in Methods for Post-training and Elicitation (The Alignment Project by UK AISI)
Ω
1mo
Ω
0
10Research Areas in Benchmark Design and Evaluation (The Alignment Project by UK AISI)
Ω
1mo
Ω
0
3Research Areas in Probabilistic Methods (The Alignment Project by UK AISI)
Ω
1mo
Ω
0
14Research Areas in Evaluation and Guarantees in Reinforcement Learning (The Alignment Project by UK AISI)
Ω
1mo
Ω
0
28The Alignment Project by UK AISI
Ω
1mo
Ω
0
57An alignment safety case sketch based on debate
Ω
4mo
Ω
21
113UK AISI’s Alignment Team: Research Agenda
Ω
4mo
Ω
2
57A sketch of an AI control safety case
Ω
7mo
Ω
0
72Automation collapse
Ω
10mo
Ω
9
7Should you work at a leading AI lab? (including in non-safety roles)
2y
0
Load More