Truthful LMs as a warm-up for aligned AGI — LessWrong