Towards Evaluating AI Systems for Moral Status Using Self-Reports — LessWrong