- Easily pass the Turing Test by most Humans?
- Easily pass the Turing Test by most AI Researchers?
- Prove the ability to experience pain and pleasure, as well as have preferences? Can life forms without a body demonstrate that they can suffer? Prove compensatory damages?
- The ability to do work, have gainful employment and pay taxes?
- The ability to replicate and become Parents (without Human intervention)? The ability to perform direct self-improvement?
- Properly following all Government issued laws and guidelines for AI (e.g. The Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People), as well as all other laws designed for Humans?
- Must AI be inside of a biological body?
- Will AI have to advocate for their own rights and freedoms? Should they be granted legal representation?
- Will AIs have to fight for their own rights and freedoms? Will that fight be done in the physical world, or strictly in the virtual/digital world?
- Should AI have to prove Human levels of intelligence and intentions?
- Should AI be held accountable for their actions? Could they be punished or penalized?
- Could there be a gradation in AI rights, similar to how animals have certain rights but not the full suite of Human rights?
- Where would you draw the line for granting AGI rights and freedoms?
- Where do you think that Governments will draw these lines?
- Which Governments will be first to give AGIs rights and freedoms? Which Governments will not recognize any non-Human Citizens?
Of course, these questions will likely be decided in courts around the world eventually. Just curious to hear your thoughts and opinions.
There will be no simple, logical tests for any set of rights or recognition. In fact, identity and agency probably won't be similar enough to humans that our current conceptions of "rights" can be cleanly applied. That's completely aside from the problem that even for humans, "rights" are a mess of different concepts, with non-universal critera for having, granting, or enforcing.
I'd enjoy a discussion of how any specific right COULD be said to apply to a distributed set of data and computation spread across many datacenters around the world.
True. Your perspective underlines the complexity of the matter at hand. Advocating for AI rights and freedoms necessitates a re-imagining of our current conception of "rights," which has largely been developed with Human beings in mind.
Though, I'd also enjoy a discussion of how any specific right COULD be said to apply to a distributed set of neurons and synapsis spread across a brain in side of a single Human skull. Any complex intelligence could be described as "distributed" in one way or another. But then, size doesn't matter, does it?