Robert Long On Why Artificial Sentience Might Matter — LessWrong