What does it mean for an LLM such as GPT to be aligned / good / positive impact? — LessWrong