I think that the real existential threat is something which could be described by one word: multipandemic.

That is many simultaneous deadly pandemics, may be organised artificially or because of quick growth of the number of bioterrorists and availability of synthetic biology. I wrote an article about it, but it needs major revision.

Hmm. I could see that being a serious threat, at least a potentially civilization-ending one.

Again though, would you agree that the best way to reduce the risk of this threat is biotech research itself?

Could the Maxipok rule have catastrophic consequences? (I argue yes.)

by philosophytorres 1 min read25th Aug 201732 comments

6


Here I argue that following the Maxipok rule could have truly catastrophic consequences.

Here I provide a comprehensive list of actual humans who expressed, often with great intensity, omnicidal urges. I also discuss the worrisome phenomenon of "latent agential risks."

And finally, here I argue that a superintelligence singleton constitutes the only mechanism that could neutralize the "threat of universal unilateralism" and the consequent breakdown of the social contract, resulting in a Hobbesian state of constant war among Earthians.

I would genuinely welcome feedback on any of these papers! The first one seems especially relevant to the good denizens of this website. :-)