LESSWRONG
LW

AI GovernanceAI RiskExistential riskAI
Personal Blog

35

New US Senate Bill on X-Risk Mitigation [Linkpost]

by Evan R. Murphy
4th Jul 2022
1 min read
12

35

This is a linkpost for https://www.hsgac.senate.gov/media/majority-media/peters-introduces-bipartisan-bill-to-ensure-federal-government-is-prepared-for-catastrophic-risks-
AI GovernanceAI RiskExistential riskAI
Personal Blog

35

New US Senate Bill on X-Risk Mitigation [Linkpost]
12Davidmanheim
11Catherine Low
-9Jeff Rose
8Davidmanheim
7GeneSmith
2Evan R. Murphy
3Shiroe
3Rob Bensinger
2Rob Bensinger
1Shiroe
3Rob Bensinger
2Daniel_Eth
New Comment
Rendering 11/12 comments, sorted by
top scoring
(show more)
Click to highlight new comments since: Today at 5:25 AM
[-]Davidmanheim3y*124

The text is now available, here: https://www.congress.gov/bill/117th-congress/senate-bill/4488

Reply
[-]Catherine Low3y*1116

This bill does seem very important. It is hard to know what will help or hinder the political process, so I recommend that folks in the EA and LW community don't try to do a public coordinated effort try to influence the content or outcome of this proposed bill - at least for now. 

My understanding is that the people involved in drafting this bill are aware of the EA and LW community, so they know they can reach out when and if they think that would be helpful.

Reply
[+]Jeff Rose3y-9-7
[-]GeneSmith3y70

How well have these types of inter-agency committees tended to work in the past? Is this a good way to actually get things done or does it just add more bureaucracy?

Reply
[-]Evan R. Murphy3y*21

Good question. I'm not sure about this types of committees in particular, but:

One reason this might not go terribly is that unlike many issues government deals with, there probably aren't a mess of competing interests they'll have to cater to in this case. Voters don't feel strongly about obscure catastrophic risks, and I can't think of any powerful companies who would be investing in lobbying around this (they mostly care about short-term issues).

So if the senators care about this issue and have good guidance on it, they will be relatively unencumbered to follow their experts' advice. They won't have to, e.g. contort their plans to sound good to their constituents and then hollow them out to please their campaign donors.

Reply
[-]Shiroe3y3-7

It's interesting the the term 'abused' was used with respect to AI. It makes me wonder if the bill has misalignment risks in mind at all or only misuse risks.

I would be very surprised if they had anything like the Yudkowskian paradigm in mind when they were thinking of this.

Reply
[-]Rob Bensinger3y33

Why? ~All the other gov stuff I'm aware of that talks about "GCR" or that talks about AI in the context of "high-consequence [catastrophic] events, regardless of the low probability" cites Bostrom, MIRI, Ord, or Stuart Russell.

Reply
[-]Rob Bensinger3y22

(But I agree they're likely to have views closer to Superintelligence, Human Compatible, or The Precipice, rather than AGI Ruin. I just think of those views as pretty close to the Yudkowskian paradigm -- eg, Bostrom is big on paperclippers and foom.)

Reply
[-]Shiroe3y10

Bostrom and MIRI being cited is pretty cool. I would have thought they'd be outside the Overton window. EDIT: Do you know when the earliest citations occurred?

Reply
[-]Rob Bensinger3y30

E.g., Preparing for the Future of Artificial Intelligence and Wired in 2016.

Reply
[-]Daniel_Eth3y20

It's interesting the term 'abused' was used with respect to AI. It makes me wonder if the authors have misalignment risks in mind at all or only misuse risks.

 

A separate press release says, "It is important that the federal government prepare for unlikely, yet catastrophic events like AI systems gone awry" (emphasis added), so my sense is they have misalignment risks in mind.

Reply
Moderation Log
More from Evan R. Murphy
View more
Curated and popular this week
11Comments

Two US Senators have introduced a bipartisan bill specifically focused on x-risk mitigation, including from AI. From the post on Senate.gov (bold mine):

WASHINGTON, D.C. – U.S. Senator Gary Peters (MI), Chairman of the Homeland Security and Governmental Affairs Committee, introduced a bipartisan bill to ensure our nation is better prepared for high-consequence events, regardless of the low probability, such as new strains of disease, biotechnology accidents, or naturally occurring risks such as super volcanoes or solar flares that though unlikely, would be exceptionally lethal if they occurred.

“Making sure our country is able to function during catastrophic events will improve national security, and help make sure people in Michigan and across the country who are affected by these incidents get the help they need from the federal government,” said Senator Peters. “Though these threats may be unlikely, they are also hard to foresee, and this bipartisan bill will help ensure our nation is prepared to address cataclysmic incidents before it’s too late.”

[The legislation] will establish an interagency committee for risk assessment that would report on the adequacy of continuity of operations (COOP) and continuity of government (COG) plans for the risks identified. The bipartisan legislation would also help counter the risk of artificial intelligence (AI), and other emerging technologies from being abused in ways that may pose a catastrophic risk.

[...]

It's interesting the term 'abused' was used with respect to AI. It makes me wonder if the authors have misalignment risks in mind at all or only misuse risks.

I haven't been able to locate the text of the bill yet. If someone finds it, please share in the comments.

Cross-posted to EA Forum. Credit to Jacques Thibodeau for posting a link on Slack that made me aware of this.