Request for feedback - infohazards in testing LLMs for causal reasoning? — LessWrong