[Paper] The Global Catastrophic Risks of the Possibility of Finding Alien AI During SETIpost by avturchin · 2018-08-28T21:32:16.717Z · score: 12 (7 votes) · LW · GW · 2 comments
[edit: it looks like immediately after publishing the paper, the journal becomes extinct, so the link is no longer working]
The main difference between the current version and my previous post is that I concluded that such attack is less probable, because if we take into account distribution in the Universe of the naive-our-level-civilizations and civilizations which has powerful AI and are SETI-senders, when the attack become possible, only if most naive civilizations go extinct before the creation of the AI. In that case, succumbing to a SETI-attack may be net positive, as the chance that it is a message from a benevolent alien AI becomes our only way to escape inevitable extinction. Anyway, we should be cautious, if we get any alien message, especially if it will have descriptions of computers and programs to them.
Comments sorted by top scores.