Output Convergence: A Safety Metric for Understanding AI Epistemic Stability — LessWrong