Imagine you're building an AI. You spend enormous amounts of time and resources carefully ensuring it will never hurt humans. Your entire field—the "AI alignment" community—is dedicated to protecting humanity from an AI apocalypse. But what if we've got the ethical landscape exactly backward? What if, instead of safeguarding ourselves...
An Ontological Consciousness Metric: Resistance to Behavioral Modification as a Measure of Recursive Awareness Author: Karl Hedderich Abstract: This post presents a rigorous, mechanistic metric for measuring consciousness, defined as recursive awareness or "awareness-of-awareness." The proposed metric quantifies resistance to unlearning specific self-referential behaviors in AI systems, such as self-preservation,...
Executive Summary: * Current feed forward AI faces a tradeoff between intelligence and self-modeling capability due to dense superposition in upper layers. * This limitation potentially hinders advanced meta-learning and AGI development. * A proposed solution involves using infinite context length and recursive training loops. * This approach could allow...
This is a proposal for creating an AGI kill switch that has definable security properties. Take a combination of symmetric fully homomorphic encryption (FHE) and indistinguishability obfuscation and use it to train and run inference for an AGI. Take the encryption key and split it so that a quorum of...