Thanks for the demo! It's a blast to play with PRISM. It feels like an authentic and genuine conversation with a reflective person. Some of the outputs are even close to profound.
It's refreshing to see that someone took the non-dual perspective serious enough to integrate it into the AI model. As well as Buddhist ethics and (possibly) epistemology. That's a non-trivial move!
It's also fascinating that the paper mentions the DMN and presumably emulates its deactivation (by some means) to generate Non-dual Worldview which is how it happens in humans. However, I could not find the details how it is implemented (have low bandwidth, so just skimmed over the text).
For those who are interested, the brain may be segregated into three networks - the Default Mode Network (DMN), the Tasking Network (TN), and the Salience Network (SN). The DMN is responsible for Self-Referential Internal Narrative (SRIN) or simply rumination about the self. Most of the time it feels unpleasant as it generates worries and fears built around constructs "self and other" and "self in time". The TN is activated when we perform a task and the brain stops ruminating. It's also responsible for the state of flow. That's why it feels more pleasant when we do a task that we like to do. The SN switches between the two.
In that model Non-dual state is a state without SRIN, where the DMN is deactivated along with "self and other" and "self in time" and the TN is activated permanently generating the flow state. So it's a state with reduced or no thoughts. It can be accomplished with some practices (and a lot of luck).
But seeing the non-dual perspective implemented in AI is something incredible!
P.S. Input (from the Sūtra on the King of Meditative Absorptions):
See that the identity of the inner is empty.
See that the external is empty as well.
The one that meditates on emptiness —
No such thing exists either.
Output:
In recognizing the emptiness of both the inner self and the external world, we encounter a profound opportunity for transformation. This realization may initially evoke discomfort, as it challenges our established notions of identity and belonging. However, through meditation and reflection on this emptiness, we can cultivate a deeper understanding of ourselves and our relationships with others. Embracing this emptiness allows us to see that our identities are not fixed but are shaped by our shared experiences and connections.
Rather than viewing emptiness as a void, we can interpret it as a catalyst for empathy and community, fostering a supportive environment where we collectively navigate the complexities of existence. By engaging in dialogue that acknowledges the fluidity of identity, we can reinforce our bonds and create a sense of belonging that transcends individual isolation. This exploration encourages us to embrace our interconnectedness, recognizing that while our identities may be transient, the relationships we build and the narratives we co-create provide a foundation for mutual growth and understanding. Thus, in contemplating emptiness, we find not only a challenge to our perceptions but also a pathway to deeper connection and shared purpose.
Somehow this has escaped comment, so I'll have a go. I write from the perspective of whether it's suitable as the value system of a superintelligence. If PRISM became the ethical operating system of a posthuman civilization born on Earth, for as long as that civilization managed to survive in the cosmos - would that be a satisfactory outcome?
My immediate thoughts are: It has a robustness, due to its multi-perspective design, that gives it some plausibility. At the same time, it's not clear to me where the seven basis worldviews come from. Why those seven, and no others? Is there some argument that these seven form a necessary and sufficient basis for ethical behavior by human-like beings and their descendants?
If I dig a little deeper into the paper, the justification is actually in part 2. Specifically, on page 12, six brain regions and their functions are singled out, as contributing to human decision-making at increasingly abstract levels (for the hierarchy, see page 15). The seven basis worldviews correspond to increasing levels of mastery of this hierarchy.
I have to say I'm impressed. I figured that the choice of worldviews would just be a product of the author's intuition, but they are actually grounded in a theory of the brain. One of the old dreams associated with CEV, was that the decision procedure for a human-friendly AI would be extrapolated in a principled way from biological facts about human cognition, rather than just from a philosophical system, hallowed tradition, or set of community principles. June Ku's MetaEthical AI, for example, is an attempt to define an algorithm for doing this. Well, this is a paper written by a human being, but the principles in part 2 are sufficiently specific, that one could actually imagine an automated process following them, and producing a form of PRISM as its candidate for CEV! I'd like @Steven Byrnes to have a look at this.
Hello all. I’d like to share a new paper, PRISM: Perspective Reasoning for Integrated Synthesis and Mediation. The central idea is to tackle AI alignment challenges by systematically representing and reconciling the range of human moral concerns rather than collapsing them into a single metric. PRISM draws on moral psychology, cognitive science, and neuroscience to propose a set of seven non-arbitrary “basis worldviews,” each hypothesized to capture a distinct dimension of human moral reasoning. The framework then uses a Pareto-inspired optimization scheme to balance these viewpoints, attempting to avoid the usual pitfalls of single-objective approaches.
Here is a concise overview:
Links:
Thanks for reading, and I hope you find the PRISM approach a useful reference in discussions of multi-perspective AI alignment.