Steven Byrnes called for a brain-like AGI research agenda, and three guys from the Hamburg EA community listened.
We are excited about Steven's five-star program 18.104.22.168 "Reverse-engineer human social instincts," and kicked off work a few weeks ago in June. We familiarized ourselves with Steven's brain-like AGI framework, and meet weekly now.
This post is an announcement and a request for feedback and collaboration.
We have a great skill fit:
We have already implemented a toy world and simple RL agent in a first iteration of Steven's framework. We build on top of the Python framework PettingZoo. Our code is in a private Github repo that we believe should stay private given the potential impact. Looking for thoughts on this.
We have collected a list of more than 60 candidate instincts from neuroscience and other sources that we can implement and experiments with.
The project website will be here: https://www.aintelope.net/ (effectively empty right now).
The project and our progress so far were presented at the Human-aligned AI Summer School in Prague on August 5th, where we got feedback about the project, brain-like AGI in general, and found the three participants who wanted to collaborate.
Implementing models and running tests is a proven way to test theories and check our understanding of the models. Better quick success/failure on something where it is too easy to build a big theory.
Specifically, we want to:
In the ideal case, the simulated agents show behavior consistent with having values like altruism or honesty.
Immediate next steps:
Good luck!! :)
Excited and happy that you are moving forward with this project. It's great to know that more paths to alignment are being actively investigated.