Does P vs NP imply a hard limit on Alignment? (A view from topology)
I’ve been trying to formalize a specific intuition about AI safety, but I’m hitting a wall with standard complexity theory definitions. I wanted to put this in front of the community to see if the geometry holds up. The standard alignment narrative assumes that a superintelligence can basically "solve" verification—that...
Jan 11