An AI alignment research agenda based on asymmetric debate and monitoring.
Epistemic Status: Personal research agenda exploring alignment approaches under assumptions of human coordination and bounded timelines. Not a comprehensive survey reflects my interests and current thinking. TL;DR: Under cooperative conditions and an assumption that we have to build ASI in relatively short timelines, we might solve alignment by building slightly-superhuman...
Apr 103