LESSWRONG
LW

Mass_Driver
3380316610
Message
Dialogue
Subscribe

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
An Activist View of AI Governance
Mikhail Samin's Shortform
Mass_Driver5d53

Part of the distinction I try to draw in my sequence is that the median person at CSET or RAND is not "in politics" at all. They're mostly researchers at think tanks, writing academic-style papers about what kinds of policies would be theoretically good for someone to adopt. Their work is somewhat more applied/concrete than the work of, e.g., a median political science professor at a state university, but not by a wide margin.

If you want political experts -- and you should -- you have to go talk to people who have worked on political campaigns, served in the government, or led advocacy organizations whose mission is to convince specific politicians to do specific things. This is not the same thing as a policy expert. 

For what it's worth, I do think OpenPhil and other large EA grantmakers should be hiring many more people. Hiring any one person too quickly is usually a mistake, but making sure that you have several job openings posted at any given time (each of which you vet carefully) is not.

Reply3
Mikhail Samin's Shortform
Mass_Driver5d90

I'm the author of the LW post being signal-boosted. I sincerely appreciate Oliver's engagement with these critiques, and I also firmly disagree with his blanket dismissal of the value of "standard practices." 

As I argue in the 7th post in the linked sequence, I think OpenPhil and others are leaving serious value on the table by not adopting some of the standard grant evaluation practices used at other philanthropies, and I don't think they can reasonably claim to have considered and rejected them -- instead the evidence strongly suggests that they're (a) mostly unaware of these practices due to not having brought in enough people with mainstream expertise, and (b) quickly deciding that anything that seems unfamiliar or uncomfortable "doesn't make sense" and can therefore be safely ignored. 

We have a lot of very smart people in the movement, as Oliver correctly points out, and general intelligence can get you pretty far in life, but Washington, DC is an intensely competitive environment that's full of other very smart people. If you try to compete here with your wits alone while not understanding how politics works, you're almost certainly going to lose.

Reply
Mainstream Grantmaking Expertise (Post 7 of 7 on AI Governance)
Mass_Driver14d30

These are important points, and I'm glad you're bringing them up!

  1. Is spending a lot of time to assess new grantmakers merely distressing (but still net positive in terms of extending your total grantmaking ability), or is it actually causing you to lose time in expectation? In other words, if you spend 40 hours recruiting and assessing candidates, does one of those candidates then go on to do 100+ hours of useful grantmaking work? Or is it more like 20 hours of useful grantmaking work?
  2. How closely connected is the shortage of people willing to be full-time grantmakers with an expectation that grantmakers will already be fluent in technical AI safety when they start work? I could imagine that people who could otherwise be working for ARC or Anthropic would be very difficult to lure away full-time, but there's an entire field of mainstream philanthropic foundations that mostly have full-time staff working on their grants. Could we hire some of those grantmakers full time to lend their general grantmaking expertise, evaluating things like budgets and org charts and performance targets, while relying on part-time advisors to provide technical expertise about the details of AI safety research? If not, why not?
  3. What do you see as the most likely or most important negative consequence if grantmakers try to offer highly competitive salaries? Is this something that your funders have literally refused to pay for, or are you worried about being criticized for it (by whom? what consequences would follow from that criticism?), or does it just generally increase team members' anxiety levels, or what exactly is the downside? I have definitely seen some of this paranoia you're talking about, so it's a real problem, but I wonder if it's worth accepting the costs associated with paying highly competitive salaries in order to attract more and better people. It's also worth noting that 'highly competitive nonprofit salaries' are still lower than 'highly competitive tech salaries,' probably by a factor of about 3. You can get top-notch grantmaking talent for much less than the price of top-notch computer engineering talent.
Reply
I'm scared.
Mass_Driver16d50

I mostly just got older and therefore calmer. I've crossed off most of the highest-priority items from my bucket list, so while I would prefer to continue living for a good long while, my personal death and/or defeat doesn't seem so catastrophically bad anymore, and to cope with the loss of civilization/humanity I read a lot of history and sci-fi and anthropology and other works that help me zoom out and see that there has already been great loss and that while I do want to spend my resources fighting to reduce the risk of that loss, it's not something I need to spend a lot of time or energy personally suffering over, especially not in advance. Worry is interest paid on trouble before it's due.

Reply2
Orphaned Policies (Post 5 of 7 on AI Governance)
Mass_Driver16d40

Interesting thoughts; thanks for sharing, and for your work at CeSIA. 

I've put some work into building coordination among US AI safety advocates, and it's been somewhat helpful, but there are limits to how much we can expect discussions about coordination to lead to unified action because different organizations have different funders, different principles, and different interests. Merely sharing information about what different groups are working on will not spontaneously cause those groups to pick a single task and pivot to supporting it.

Reply1
Political Funding Expertise (Post 6 of 7 on AI Governance)
Mass_Driver16d*20

I suppose I was speaking too loosely -- thank you for flagging that!

I don't mean that it's literally impossible to assess whether AI governance grants have been successful -- only that doing so requires somewhat more deliberate effort than it does for most other types of grants, and that there is relatively less in the way of established infrastructure to support such measurements in the field of AI governance. 

If you run an anti-malaria program, there's a consensus about at least the broad strokes of what you're supposed to measure (i.e., malaria cases), and you'll get at least some useful information about that metric just from running your program and honestly recording what your program officers observe as they deliver medication. If your bed nets are radically reducing the incidence of malaria in your target population, then the people distributing those bed nets will probably notice. There is also an established literature on "experimental methods" for these kinds of interventions that tells us that we need to be taking measurements and how to do so and how to interpret them.

By contrast, if you're slightly reducing the odds of an AI catastrophe, it's not immediately obvious or agreed-upon what observable changes this ought to produce in the real world, and a grant funder isn't very likely to notice those changes unless they specifically go and look for them. They're also less likely to specifically go and look for them in an effective way, because the literature on experimental methods for politics is much less well-developed than the literature on experimental methods for public health.

My work so far has mostly been about doing the advocacy, rather than establishing better metrics to evaluate the impact of that advocacy. That said, in posts 1 and 7 of this sequence, I do suggest some starting points. I encourage funders to look at figures like the number of meetings had with politicians, the number of events that draw in a significant number of politicians, the number of (positive) mentions in mainstream 'earned media', the number of endorsements that are included in Congressional offices' press releases, and the number (and relative importance) of edits made to Congressional bills.

If your work is focused on the executive or judicial branch instead of on Congress, you could adapt some of those metrics accordingly, e.g., edits to pending regulation or executive orders, or citations to your amicus curiae briefs in judicial opinions, and so on.

Reply11
So You Want to Work at a Frontier AI Lab
Mass_Driver25d41

> frontier labs are only pretending to try to solve alignment 

>>This is probably the main driver of our disagreement.

I agree with your diagnosis! I think Sam Altman is a sociopathic liar, so the fact that he signed the statement on AI risk doesn't convince me that he cares about alignment. I feel reasonably confident about that belief. Zvi's series on Moral Mazes apply here: I don't claim that you literally can't mention existential risk at OpenAI, but if you show signs of being earnestly concerned enough about it to interefere with corporate goals, then I believe you'll be sidelined.

I'm much less confident about whether or not successful alignment looks like normal deep learning work; I know more about corporate behavior than I do about technical AI safety. It seems odd and unlikely to me that the same kind of work (normal deep learning) that looks like it causes a series of major problems (power-seeking, black boxes, emergent goals) when you do a moderate amount of it would wind up solving all of those same problems when you do a lot of it, but I'm not enough of a technical expert to be sure that that's wrong.

Because there are independent, non-technical reasons for people to want to believe that normal deep learning will solve alignment (it means they get to take fun, high-pay, high-status jobs at AI developers without feeling guilty about it), if you show me a random person who believes this and I don't know anything about their incorruptiability or the clarity of their thinking ahead of time, then my prior is that most of the people in the random distribution that this person was drawn from probably arrived at the belief mostly out of convenience and temptation, rather than mostly by becoming technically convinced of the merits of a position that seems a priori unlikely to me. However, I can't be sure -- perhaps it's more likely than I think that normal deep learning can solve alignment.

Reply
So You Want to Work at a Frontier AI Lab
Mass_Driver25d41

Well, I can't change the headline; I'm just a commenter. However, I think the reason why "frontier labs will fail at alignment while nonprofits can succeed" is that frontier labs are only pretending to try to solve alignment -- it's not actually a serious goal of their leadership, and it's not likely to get meaningful support in terms of compute, recruiting, data, or interdepartmental collaboration, and in fact the leadership will probably actively interfere with your work on a regular basis because the intermediate conclusions you're reaching will get in the way of their profits and hurt their PR. In order to do useful superalignment research, I suspect you sometimes need to warn about or at least openly discuss the serious threats that are posed by increasingly advanced AI, but the business model of frontier labs depends on pretending that none of those threats are actually serious. By contrast, the main obstacle at a nonprofit is that they might not have much funding, but at least whatever funding they do have will be earnestly directed at supporting your team's work.

Reply
So You Want to Work at a Frontier AI Lab
Mass_Driver1mo41

I suspect Joe would agree with me that the current odds that AI developers solve superalignment are significantly less than 20%.

Even if we concede your estimate of 20% for the sake of argument, though, what price are you likely to pay for increasing the odds of success by 0.01%? Suppose that, given enough time, nonprofit alignment researchers would eventually solve superalignment with 80% odds. In order to increase, e.g., Anthropic's odds of success by 0.01%, are you boosting Anthropic's capabilities in a way that shortens timelines in a way that decreases the amount of time that the nonprofit alignment teams have to solve superalignment in a way that reduces their odds of success by at least 0.0025%? If so, you've done net harm. If not, why not? What about Joe's arguments that most for-profit alignment work has at least some applicability to capabilities do you find unconvincing?

Reply
Orphaned Policies (Post 5 of 7 on AI Governance)
Mass_Driver1mo20

Transparency is less neglected than some other topics -- check out HR 5539 (Transparent by Design Act), S 3312 (AIRIA), and HR 6881 (AI Foundation Model Transparency Act).

There's room for a little bit more useful drafting work here, but I wouldn't call it orphaned, exactly.

Reply
Load More
43Mainstream Grantmaking Expertise (Post 7 of 7 on AI Governance)
22d
5
50Political Funding Expertise (Post 6 of 7 on AI Governance)
26d
4
64Orphaned Policies (Post 5 of 7 on AI Governance)
2mo
5
55Shift Resources to Advocacy Now (Post 4 of 7 on AI Governance)
2mo
18
109We're Not Advertising Enough (Post 3 of 7 on AI Governance)
2mo
10
54The Need for Political Advertising (Post 2 of 7 on AI Governance)
2mo
2
113Please Donate to CAIP (Post 1 of 7 on AI Governance)
2mo
20
10LINK: Quora brainstorms strategies for containing AI risk
9y
1
18Help Build a Landing Page for Existential Risk?
10y
32
-20Reminder Memes
12y
4
Load More