To what ethics is an AGI actually safely alignable? — LessWrong