rorygreig

Research Engineer in AI Safety

Wiki Contributions

Comments

Thanks again John for giving this talk! I really enjoyed the talk at the time and was pleasantly surprised with the positive engagement from the audience. I'm also pleased that this turned into a resource that can be re-shared.

I suppose I can imagine an architecture that has something like a central planning agent that is capable of having a goal, observing the state of the world to check if the goal had been met, coming up with high level strategies to meet that goal, then delegating subtasks to a set of subordinate sub-agents (whilst making sure that these tasks are broken down enough that the sub-agents themselves don't have to do much long time-horizon planning or goal directed behaviour).

With this architecture it seems like all the agent-y goal-directed stuff is done by a single central agent.

However I do agree that this may be less efficient or capable in practice than an architecture with more autonomous, decentralised sub-agents. But on the other hand it might be better at more consistently pursuing a stable goal, so that could compensate.

Good points, however I'm still a bit confused about the difference between two different scenarios: "multiple sub-agents" vs "a single sub-agent that can use tools" (or can use oracle sub-agents that don't have their own goals).

For example a human doing protein folding using alpha-fold; I don't think of that as multiple sub-agents, just a single agent using an AI tool for a specialised task (protein folding). (Assuming for now that we can treat a human as a single agent, which isn't really the case, but you can imagine a coherent agent using alpha-fold as a tool).

It still seems plausible to me that you might have a mind made of many different parts, but there is a clear "agent" bit that actually has goals and is controlling all the other parts.

I agree that initially a powerful AGI would likely be composed of many sub-agents. However it seems plausible to me that these sub-agents may “cohere” under sufficient optimisation or training. This could result in the sub-agent with the most stable goals winning out. It’s possible that strong evolutionary pressure makes this more likely.

You could also imagine powerful agents that aren’t composed of sub-agents, for example a simpler agent with very computationally expensive search over actions.

Overall this topic seems under-discussed in my opinion. It would be great to have a better understanding of whether we expect sub-agents to turn into a single coherent agent.

I really enjoyed this dialogue, thanks!

A few points on complexity economics:

The main benefit of complexity economics in my opinion is that it addresses some of the seriously flawed and over-simplified assumptions that go into classical macroenomic models, such as rational expectations, homogenous agents, and that the economy is at equilibrium.  However it turns out that replacing these with more relaxed assumptions is very difficult in practice.  Approaches such as agent-based models (ABMs) are tricky to get right, since they have so many degrees of freedom. However I do think that this is a promising avenue of research, but it maybe it still needs more time and effort to pay off.  Although it's possible that I'm falling into a "real communism has never been tried" trap.

I also think that ML approaches are very complementary to simulation based approaches like ABMs.

In particular the complexity economics approach is useful for dealing with the interactions between the economy and other complex systems, such as public health. There was some decent research done on economics and the covid pandemic, such as Doyne Farmer's work: https://www.doynefarmer.com/covid19-research, who is a well known complexity scientist.

It's hard to know how much of this "heterodox" economics would have happened anyway, even in the absence of people who call themselves complexity scientists. But I do think complexity economics played a key role in advocating for these new approaches.

Having said that: I'm not an economist, so I'm not that well placed to criticise the field of economics.

More broadly I found the discussion on self-referential and recursive predictions very interesting, but I don't necessarily think of that as central to complexity science.

I'd also be interested in hearing more about how this fits in with AI Alignment, in particular complexity science approaches to AI Governance.

The workshop talks from the previous year's ALIFE conference (2022) seem to be published on YouTube, so I'm following up with whether John's talk from this year's conference can be released as well.

This is a really interesting point that I hadn't thought of!

I'm not sure where I land on the conclusion though. My intuition is that two copies of the same mind emulation running simultaneously (assuming they are both deterministic and are therefore doing identical computations) would have more moral value than only a single copy, but I don't have a lot of confidence in that. 

Yes it is indeed a hybrid event! 

I have now added the following text to the website:

The conference is hybrid in-person / virtual. All sessions will have remote dial-in facilities, so authors are able to present virtually and do not need to attend in-person.

This was in our draft copy for the website, I could have sworn it was on there but somehow it got missed out, my apologies!

Load More