[edit: it looks like immediately after publishing the paper, the journal becomes extinct, so the link is no longer working]

My article on the topic has been finally published, 10 years after first draft. I have discussed the problem before on LW. The preprint, free of paywall, is here.

The main difference between the current version and my previous post is that I concluded that such attack is less probable, because if we take into account distribution in the Universe of the naive-our-level-civilizations and civilizations which has powerful AI and are SETI-senders, when the attack become possible, only if most naive civilizations go extinct before the creation of the AI. In that case, succumbing to a SETI-attack may be net positive, as the chance that it is a message from a benevolent alien AI becomes our only way to escape inevitable extinction. Anyway, we should be cautious, if we get any alien message, especially if it will have descriptions of computers and programs to them.

New Comment
2 comments, sorted by Click to highlight new comments since: Today at 11:20 PM

Just out of curiosity: How probable do you think any SETI-contact will turn out to be AI-initiated as opposed to biological (in the broadest possible sense of that word)?

My estimations:

P(SETI) =0.01


New to LessWrong?