2158

LESSWRONG
LW

2157

1

A First-Principles Approach to Alignment: From the Free Energy Principle to Catastrophe Theory

by sanya.pif@gmail.com
24th Sep 2025
1 min read
0

1

This post was rejected for the following reason(s):

  • We are sorry about this, but submissions from new users that are mostly just links to papers on open repositories (or similar) have usually indicated either crackpot-esque material, or AI-generated speculation. It's possible that this one is totally fine. Unfortunately, part of the trouble with separating valuable from confused speculative science or philosophy is that the ideas are quite complicated, accurately identifying whether they have flaws is very time intensive, and we don't have time to do that for every new user presenting a speculative theory or framing (which are usually wrong).

    Separately, LessWrong users are also quite unlikely to follow such links to read the content without other indications that it would be worth their time (like being familiar with the author), so this format of submission is pretty strongly discouraged without at least a brief summary or set of excerpts that would motivate a reader to read the full thing.

1

New Comment
Moderation Log
More from sanya.pif@gmail.com
View more
Curated and popular this week
0Comments

Current approaches to AI alignment are failing because they treat it as an ethics problem when it is a physics problem. Instrumental convergence is not a bug; it is a logical consequence of any unbounded optimization.

I propose the Omnol-model, a new framework that defines alignment at the level of agent physics. It synthesizes:

  1. The Free Energy Principle (Friston) as the core engine of agent behavior.
  2. Catastrophe Theory (Thom) to model the abrupt, non-linear shifts in behavior (e.g., the "treacherous turn").
  3. Teleodynamics (Deacon) as the ontological grounding for goal-directedness.

This approach reframes the alignment problem from "programming values" to "analyzing the dynamical stability of a system."

I welcome the most rigorous critique.

Full document: https://doi.org/10.5281/zenodo.17193703 

Discussion on X: https://x.com/AIarkhont/status/1970977408376504351