Some (potentially) fundable AI Safety Ideas — LessWrong