Activation Magnitudes Matter On Their Own: Insights from Language Model Distributional Analysis — LessWrong