Rejected for the following reason(s):
- Insufficient Quality for AI Content.
- We are sorry about this, but submissions from new users that are mostly just links to papers on open repositories (or similar) have usually indicated either crackpot-esque material, or AI-generated speculation.
- Difficult to evaluate, with potential yellow flags.
Read full explanation
I have been exploring this question for quite some time now and recently published a preprint proposing a nested VSM architecture for the purpose. Coming from a systems/cybernetics background, honestly, it received a traction I haven't expected, so I pulled together the courage to share it with the "hardcore alignment people" here on this platform.
My core hypothesis is that we replace pure reward maximization with what may be called viability maximization. My research questions were simple:
I have provided algorithmic sketches and experimental protocols for those interested: https://zenodo.org/records/17943102
I would appreciate critique on the formalization of the coherence metric and the control topology.