Reinforcement Learning: Essential Step Towards AGI or Irrelevant? — LessWrong