Human Alignment

Applied to Antagonistic AI by Xybermancer 2mo ago
Applied to 3. Uploading by RogerDearnaley 5mo ago

Human alignment is a state of humanity in which most or all of humanity systematically cooperates to achieve positive-sum outcomes for everyone (or at a minimum are prevented from pursuing negative sum outcomes), in a way perpetually sustainable into the future. Such a state of human alignment may be necessary to prevent an existential catastrophe in the case that the "Vulnerable"Vulnerable World Hypothesis" is correct.

Human alignment is a state of humanity in which most or all of humanity systematically cooperates to achieve positive-sum outcomes for everyone (or at a minimum are prevented from pursuing negative sum outcomes), in a way perpetually sustainable into the future. Such a state of human alignment may be necessary to prevent an existential catastrophe in the case that the "Vulnerable World Hypothesis" is correct.

Created by Jordan Arel at 1y