What actual bad outcome has "ethics-based" RLHF AI Alignment already prevented? — LessWrong