Request for Comment
I am currently formalizing the syntax for the Coherence_Check function and defining the minimal ontology required for general-purpose LLM guardrails.
Abstract
Current alignment techniques, such as Reinforcement Learning from Human Feedback (RLHF), primarily address safety as a probabilistic reward function. While effective for style and tone, this approach fails to prevent "hallucinations" in out-of-distribution (OOD) scenarios because it lacks a mechanism for runtime verification.
This proposal introduces Relational Coherence Theory (RCT) as a neuro-symbolic architectural layer. Unlike standard "Chain of Thought" prompting, RCT functions as a deterministic logic gate. It introduces a recursive verification step—specifically, an "Ontological Guardrail"—that validates the logical coherence of relationships between subject and object before token generation.
We argue that... (read 963 more words →)