Probing Power-Seeking in LLMs — LessWrong