Prospects for Alignment Automation: Interpretability Case Study — LessWrong