simeon_c

@SaferAI

Wiki Contributions

Comments

simeon_c1411

Thanks for answering, that's very useful. 

My concern is that as far as I understand, a decent number of safety researchers are thinking that policy is the most important area, but because, as you mentioned, they aren't policy experts and don't really know what's going on, they just assume that Anthropic policy work is way better than those actually working in policy judge it to be. I've heard from a surprisingly high number of people among the orgs that are doing the best AI policy work that Anthropic policy is mostly anti-helpful. 

Somehow though, internal employees keep deferring to their policy team and don't update on that part/take their beliefs seriously. 

I'd generally bet Anthropic will push more for policies I personally support than any other lab, even if they may not push as much as I want them to.

If it's true, it is probably true to an epsilon degree, and it might be wrong because of weird preferences of a non-safety industry actor.  AFAIK,  Anthropic has been pushing against all the AI regulation proposals to date. I've still to hear a positive example.

simeon_c912

How aware were you (as an employee) & are you (now) of their policy work? In a world model where policy is the most important stuff, it seems to me like it could tarnish very negatively Anthropic's net impact.

This is the best alignment plan I've heard in a while.

simeon_c140

You are a LessWrong reader, want to push humanity's wisdom and don't know how to do so? Here's a workflow: 

  1. Pick an important topic where the entire world is confused 
  2. Post plausible sounding takes with a confident tone on it
  3. Wait for Gwern's comment on your post 
  4. Problem solved

See an application of the workflow here: https://www.lesswrong.com/posts/epgCXiv3Yy3qgcsys/you-can-t-predict-a-game-of-pinball?commentId=wjLFhiWWacByqyu6a

Playing catch-up is way easier than pushing the frontier of LLM research. One is about guessing which path others took, the other one is about carving a path among all the possible ideas that could work.

If China stopped having access to US LLM secrets and had to push the LLM frontier rather than playing catch up, how slower would it be at doing so? 

My guess is at least >2x and probably more but I'd be curious to get takes. 

simeon_c4352

Great initiative! Thanks for leading the charge on this.

Jack Clark: “Pre-deployment testing is a nice idea but very difficult to implement,” from https://www.politico.eu/article/rishi-sunak-ai-testing-tech-ai-safety-institute/

Thanks for the answer it makes sense.

To be clear I saw it thanks to Matt who did this tweet so credit goes to him: https://x.com/SpacedOutMatt/status/1794360084174410104?t=uBR_TnwIGpjd-y7LqeLTMw&s=19

simeon_c2015

Thanks for sharing. It's both disturbing from a moral perspective and fascinating to read.

Load More