Using Threats to Achieve Socially Optimal Outcomes — LessWrong