230

LESSWRONG
LW

229
AI RiskAIWorld Optimization
Frontpage

14

Taboo P(doom)

by [anonymous]
3rd Feb 2023
1 min read
10

14

14

Taboo P(doom)
6steven0461
6Vladimir_Nesov
6Dave92F1
1[anonymous]
3Shmi
2the gears to ascension
2Tristan Cook
2Thomas
1Thomas Kwa
2the gears to ascension
New Comment
10 comments, sorted by
top scoring
Click to highlight new comments since: Today at 2:15 PM
[-]steven04613y63

There's another issue where "P(doom)" can be read either as the probability that a bad outcome will happen, or the probability that a bad outcome is inevitable. I think the former is usually what's meant, but if "P(doom)" means "the probability that we're doomed", then that suggests the latter as a distracting alternative interpretation.

Reply
[-]Vladimir_Nesov3y60

I think it's Bostrom's notion of existential catastrophe that introduced the conflation of extinction with losing control over most of the cosmic endowment. The "curtail long-term potential" term is ambiguous between things like making current technological level unachievable (with something like a civilization-decimating pandemic that somehow can be survived but never truly recovered from), and making hypothetical CEV-guided capture of future lightcome unachievable, while still granting something like CEV-guided capture of merely our galaxy.

This was cemented in the common use by arguments about AI risk that boiled down to lack of the full CEV-guided future lightcone almost certainly implying extinction, so that there is little use for the category of outcomes somewhere in the middle. But recently there are possibilities like even significantly-alien-on-reflection LLMs capturing the cosmic endowment for themselves without exterminating humanity and possibly even gifting some post-singularity boons like uploading. This makes the non-extinction existential catastrophe a plausible outcome, one much preferable to an extinction existential catastrophe (which still seems likely with even more alien AGI designs that are not far behind, or with LLMs themselves failing the AI risk resistance check by building unaligned AGIs).

Reply
[-]Dave92F13y60

Just for the record, I think there are two important and distinguishable P(doom)s, but not the same two as NathanBarnard:

P(Doom1): Literally everyone dies. We are replaced by either by dumb machines with no moral value (paperclip maximisers) or by nothing.

P(Doom2): Literally everyone dies. We are replaced by machines with moral value (conscious machines?), who go on to expand a rich culture into the universe.

Doom1 is cosmic tragedy - all known intelligence and consciousness are snuffed out. There may not be any other elsewhere, so potentially forever.

Doom2 is maybe not so bad. We all die, but we were all going to die anyway, eventually, and lots of us die without descendants to carry our genes, and we don't think that outcome is so tragic. Consciousness and intelligence spreads thru the universe. It's a lot like what happened to our primate ancestors, before Homo sapiens. In some sense the machines are our descendants (if only intellectual) and carry on the enlightening of the universe.

Reply
[-][anonymous]3y10

Oh yes, this is also a very important distinction (although I would only value outcome 2 if the machines were conscious and living good lives.) 

Reply
[-]Shmi3y32

It's more like "Taboo Doom".

Reply
[-]the gears to ascension3y20

Some related work:

  • via google "against p(doom) site:lesswrong.com"  ->
    • exact match https://www.lesswrong.com/posts/Tt83WGrZn6QGcnf2B/against-a-general-factor-of-doom 
    • https://www.lesswrong.com/posts/tL3QsPfrR83JKLSXy/why-i-don-t-believe-in-doom 
    • https://www.lesswrong.com/posts/Tt83WGrZn6QGcnf2B/against-a-general-factor-of-doom 
    • https://www.lesswrong.com/posts/sAEE7fdnv3KpcaQEi/when-people-ask-for-your-p-doom-do-you-give-them-your-inside 
  • soft matches via similarity search using metaphor greaterwrong version of this taboo P(doom) post ->
    •  https://www.lesswrong.com/posts/LDRQ5Zfqwi8GjzPYG/counterarguments-to-the-basic-ai-x-risk-case 
    • https://www.lesswrong.com/posts/9MZdwQ7u53oaRiBYX/against-aligning-with-neo-luddites 
    •  
Reply
[-]Tristan Cook3y20

I agree. I think we should break "doom" into at least these four outcomes {human extinction, humans remain on Earth} x {lots of utility achieved, little to no utility} ( {E,¬E}×{U,¬U})

Reply
[-]Thomas3y21

Upvoted for a fresh, non-forced by an ultra-utilitaristic POV. With this approach, p(Giga-doom) is also much lower, I guess.

Reply
[-]Thomas Kwa3y1-3

I like this point, but downvoted this because it didn't reference prior work

Reply
[-]the gears to ascension3y20

Somewhat reasonable, but I'd argue that it's preferable to do the citations lit review as a commenter. eg, I'll give it a shot.

Reply
Moderation Log
Curated and popular this week
10Comments
AI RiskAIWorld Optimization
Frontpage
Mentioned in
42Stop talking about p(doom)

I think it's actively unhelpful to talk about P(doom) because it fails to distinguish between literally everyone dying and humanity failing to capture 99.99% of the value of the future under a total utilitarian view but in practice, everyone who's born lives a very good life. These are very different outcomes and it's unhelpful not to distinguish between them and everything else in that spectrum. 

This is especially the case since astronomical waste arguments really only bite for total utilitarian views. Under moral views where potential people not coming into existence is more similar to preventing someone from living an extraordinarily happy life rather than a merely happy life, as opposed to preventing someone coming into existence being similar to murder, it's quite reasonable to prioritise other goals well above preventing astronomical waste. Under these non-totalalist views preventing totalitarian lock-in or S-risks might look much more important than ensuring we don't create 10^(very large number) of happy lives. 

I think this also matters on a practical level when talking about threat models of AI risks. Two people could have the same p(doom) but one is talking about humans being stripped for their atoms and the other is talking about slow disempowerment in which no one actually dies and everyone, in fact, could be living very good lives but humanity isn't able to capture almost all of the value of future from a total utilitarian perspective. These plausibly require different interventions to stop them from happening. 

It also seems like one's prior on humanity going extinct as a result of AI should be quite different from disempowerment, but people often talk about what their prior on P(doom) should be as a univariate probability distribution.