I just saw Jan Kulveit's Announcement of the Alignment of Complex Systems Research Group, and was very excited to see other people in the Alignment Space noticing a need for a formal theory of alignment of agents arranged in a hierarchy. I also think formalizing how agents can work together to create an agent at a higher abstraction level to solve some collective need of the subsystems is the very first step towards understanding how we can hope to align powerful AGIs in different contexts.
Some thoughts on this:
I think selecting the right upward reward mechanism is an important thing for picking AGI research and engineering methodology; since we probably want an alignment mechanism that is robust to recursive self-improvement take-off scenarios, evolution would probably not be a good reward mechanism, for example (As by the time the selection would be supposed to happen, it would be too late.).
I also want to note that I took the notion of subsystem alignment from an interview of Joscha Bach.