What Should AI Owe To Us? Accountable and Aligned AI Systems via Contractualist AI Alignment
This is an extended and edited transcript of the talk I recently gave at EAGxSingapore 2022. The title has been changed for easier searchability of "Contractualist AI Alignment". Abstract: Artificial intelligence seems poised to alter our civilization in transformative ways. How can we align this development with our collective interests?...


It seems to me that it's not right to assume the probability of opportunities to trade are zero?
Suppose both John and David are alive on a desert island right now (but slowly dying), and there's a chance that a rescue boat will arrive that will save only one of them, leaving the other to die. What would they contract to? Assuming no altruistic preferences, presumably neither would agree to only the other person being rescued.
It seems more likely here that bargaining will break down, and one of them will kill off the other, resulting in an arbitrary resolution of who ends up on the rescue boat, not a "rational" resolution.