I have links to old LW posts in some articles and other places. What will happen with all these links?
It should be able to understand human language or it is (most likely) not dangerous.
I am getting this comment in different forms many times, so I will state again 3 main reasons why it is not exactly true:
1) The message is addressed to Young AI, not to full-blown superintelligence. Such Young AI will be time constraint but could have to make some pivotal act.
2) The message incl...(read more)
Why? if we have fullblown papercliper, which is coming of age, we are already doomed, and starting its bullshit detector will not make our situation any worse.
In fact, I am not describing here how to build such a trap, but I express our commitment to build the trap, in order to install uncertainty into any future AI if the trap exists and what it could be.
I think that real constructors of the trap will find the correct ways to make it effective and bal...(read more)
It will kill humanity not because it will be annoyed, but for two main goals: its own safety, or to use human atoms.
Other variants also possible, I explored them here: http://lesswrong.com/lw/mgf/a_map_agi_failures_modes_and_levels/
What if we fail to make them in time, and instead fooming non-human AI appear?
Surely, most arguments are weak, but it is not a problem, as we try all possible independent arguments, hoping that at least one will work. So, there is no reason to ignore putting weak arguments in the text of the message.
However, exactly this argument could be made stronger. It is intended to pr...(read more)
Another cash level is drawing visual map and when use visual field as a cash of all the elements of the problem.
Thanks! It is nice that we come to the wording "young AI" more or less independently. I tried to maximally stealmanning this approach and collect all ideas about messaging to the future AI. I will add your post in the bibliographic section.