LESSWRONG
LW

117
ryan_greenblatt
19553Ω47804718508
Message
Dialogue
Subscribe

I'm the chief scientist at Redwood Research.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
15ryan_greenblatt's Shortform
Ω
2y
Ω
314
ryan_greenblatt's Shortform
ryan_greenblatt1dΩ220
  • Anthropic reports SWE bench scores without reasoning which is some evidence it doesn't help (much) on this sort of task. (See e.g. the release blog post for 4 opus)
  • Anecdotal evidence

Probably it would be more accurate to say "doesn't seem to help much while it helps a lot for openai models".

Reply
ryan_greenblatt's Shortform
ryan_greenblatt3dΩ19320

Some people seem to think my timelines have shifted a bunch while they've only moderately changed.

Relative to my views at the start of 2025, my median (50th percentile) for AIs fully automating AI R&D was pushed back by around 2 years—from something like Jan 2032 to Jan 2034. My 25th percentile has shifted similarly (though perhaps more importantly) from maybe July 2028 to July 2030. Obviously, my numbers aren't fully precise and vary some over time. (E.g., I'm not sure I would have quoted these exact numbers for this exact milestone at the start of the year; these numbers for the start of the year are partially reverse engineered from this comment.)

Fully automating AI R&D is a pretty high milestone; my current numbers for something like "AIs accelerate AI R&D as much as what would happen if employees ran 10x faster (e.g. by ~fully automating research engineering and some other tasks)" are probably 50th percentile Jan 2032 and 25th percentile Jan 2029.[1]

I'm partially posting this so there is a record of my views; I think it's somewhat interesting to observe this over time. (That said, I don't want to anchor myself, which does seem like a serious downside. I should slide around a bunch and be somewhat incoherent if I'm updating as much as I should: my past views are always going to be somewhat obviously confused from the perspective of my current self.)

While I'm giving these numbers, note that I think Precise AGI timelines don't matter that much.


  1. See this comment for the numbers I would have given for this milestone at the start of the year. ↩︎

Reply
ryan_greenblatt's Shortform
ryan_greenblatt3dΩ330

I've updated towards somewhat longer timelines again over the last 5 months. Maybe my 50th percentile for this milestone is now Jan 2032.

Reply
AIs will greatly change engineering in AI companies well before AGI
ryan_greenblatt3dΩ460

Some AI company employees with shorter timelines than me mostly. I also think that "why I don't agree with X" is a good prompt to express some deeper aspect of my models/views. It also makes a good reasonably engaging hook for a blog post.

I might write some posts responding arguments for longer timelines that I disagree with if I feel like I have something interesting to say.

Reply
peterbarnett's Shortform
ryan_greenblatt7d40

Further discussion by Carl is here.

Reply1
Interpretability is the best path to alignment
ryan_greenblatt8d30

See also To be legible, evidence of misalignment probably has to be behavioral.

Reply
ryan_greenblatt's Shortform
ryan_greenblatt9d20

This is only somewhat related to what you were saying, but I do think 100 year medians vs 10 year medians does matter a bunch.

Reply
Anthropic's leading researchers acted as moderate accelerationists
ryan_greenblatt9d4429

The Time article is materially wrong about a bunch of stuff

Agreed which is why I noted this in my comment.[1] I think it's a bad sign that Anthropic seemingly actively sought out an article that ended up being wrong/misleading in a way which was convenient for Anthropic at the time and then didn't correct it.

I really don't want to get into pedantic details, but there's no "supposed to" time for LTBT board appointments, I think you're counting from the first day they were legally able to appoint someone. Also https://www.anthropic.com/company lists five board members out of five seats, and four Trustees out of a maximum five. IMO it's fine to take a few months to make sure you've found the right person!

First, I agree that there isn't a "supposed to" time, my wording here was sloppy, sorry about that.

My understanding was a that there was a long delay (e.g. much longer than a few months) between the LTBT being able to appoint a board member and actually appointing such a member and a long time where the LTBT only had 3 members. I think this long of a delay is somewhat concerning.

My understanding is that the LTBT could still decide one more seat (so that it determines a majority of the board). (Or maybe appoint 2 additional seats?) And that it has been able to do this for almost a year at this point. Maybe the LTBT thinks the current board composition is good such that appointments aren't needed, but the lack of any external AI safety expertise on the board or LTBT concerns me...

More broadly, the corporate governance discussions (not just about Anthropic) I see on LessWrong and in the EA community are very deeply frustrating, because almost nobody seems to understand how these structures normally function or why they're designed that way or the failure modes that occur in practise. Personally, I spent about a decade serving on nonprofit boards, oversight committes which appointed nonprofit boards, and set up the goverance for a for-profit company I founded.

I certainly don't have particular expertise in corporate governance and I'd be interested in whether corporate governance experts who are unconflicted and very familiar with the AI situation think that the LTBT has the de facto power needed to govern the company through transformative AI. (And whether the public evidence should make me much less concerned about the LTBT than I would be about the OpenAI board.)

My view is that the normal functioning of a structure like the LTBT or a board would be dramatically insufficient for governing transformative AI (boards normally have a much weaker function in practice than the ostensible purposes of the LTBT and the Anthropic board), so I'm not very satisfied by "the LTBT is behaving how a body of this sort would/should normally behave".


  1. I said something weaker: "For what it's worth, I think the implication of the article is wrong and the LTBT actually has very strong de jure power", because I didn't see anything which is literally false as stated as opposed to being misleading. But you'd know better. ↩︎

Reply
peterbarnett's Shortform
ryan_greenblatt9d25-9

People seem to be reacting to this as though it is bad news. Why? I'd guess the net harm caused by these investments is negligible and this seems like a reasonable earning to give strategy.

Reply
Trust me bro, just one more RL scale up, this one will be the real scale up with the good environments, the actually legit one, trust me bro
ryan_greenblatt9d71

Agentic software engineering mostly, I don't think Genie matters.

Reply1
Load More
45AIs will greatly change engineering in AI companies well before AGI
Ω
4d
Ω
9
151Trust me bro, just one more RL scale up, this one will be the real scale up with the good environments, the actually legit one, trust me bro
Ω
10d
Ω
26
98Attaching requirements to model releases has serious downsides (relative to a different deadline for these requirements)
Ω
17d
Ω
2
162My AGI timeline updates from GPT-5 (and 2025 so far)
Ω
24d
Ω
14
91Recent Redwood Research project proposals
Ω
2mo
Ω
0
69Jankily controlling superintelligence
Ω
3mo
Ω
4
55What does 10x-ing effective compute get you?
3mo
10
32Prefix cache untrusted monitors: a method to apply after you catch your AI
Ω
3mo
Ω
1
61AI safety techniques leveraging distillation
Ω
3mo
Ω
0
71When does training a model change its goals?
Ω
3mo
Ω
2
Load More
Anthropic (org)
8 months ago
(+17/-146)
Frontier AI Companies
a year ago
Frontier AI Companies
a year ago
(+119/-44)
Deceptive Alignment
2 years ago
(+15/-10)
Deceptive Alignment
2 years ago
(+53)
Vote Strength
2 years ago
(+35)
Holden Karnofsky
2 years ago
(+151/-7)
Squiggle Maximizer (formerly "Paperclip maximizer")
2 years ago
(+316/-20)