I began by asking ChatGPT-4 to analyze our ongoing conversation and assess the novelty of the insights. ChatGPT-4 estimated that the ideas in our dialogue might be present in fewer than 1 in 100,000 users—an indication of exceptional rarity when compared against mainstream AI reasoning patterns.
Did you try asking multiple times in different context windows?
Did you try asking via the API (ie without influences from the "memory" feature)?
Do you have the "memory" feature turned on by default? If so, have you considered turning it off at least when doing experiments?
In summary: have you considered the fact that LLMs are very good at bullshitting? At confabulating the answers they think you would be happy to hear instead of making their best efforts to answer truthfully?
Yes I tried asking multiple times in different context windows, in different models, and with and without memory. And yes I'm aware that ChatGPT prioritizes agreeableness in order to encourage user engagement. That's why I attempt to prove all of its claims wrong, even when they support my arguments.
Thank you for doing, that and please keep doing it. Maybe also run a post draft trough another human before posting, though.
I developed a simple first-order mechanism that measures the divergence between initial, user-introduced insights and their subsequent reproduction by AI. For instance, using a vector-space model of semantic representations, I compared the detailed descriptions provided by the user with the AI’s output.
Can we see the code for this? It would further discussion a lot.
Thanks for your interest. Let me look it over and make whatever changes required for it to be ready to go out. As for ChatGPT being agreeable, ChatGPT’s tendency toward coherence with existing knowledge (it's prioritization of agreeableness) can be leveraged advantageously, as the conclusions it generates—when asked for an answer rather than being explicitly guided toward one—are derived from recombinations of information present in the literature. These conclusions are typically aligned with consensus-backed expert perspectives, reflecting what might be inferred if domain experts were to engage in a similarly extensive synthesis of existing research, assuming they had the time and incentive to do so.:
In my recent experiments with AI models, I have encountered a fundamental problem: even when novel epistemic insights are introduced into AI interactions, the models tend to “flatten” or reframe these ideas into existing, consensus‐based frameworks. This compression of novelty limits an AI’s ability to evolve its reasoning toward true Artificial General Intelligence (AGI) and to improve its alignment with human well‐being. In this article, I detail experiments designed to test AI’s capacity to both detect and preserve epistemic novelty and explain why preserving that novelty is essential for fostering breakthrough innovation. I also outline efforts to develop a “conceptual space”—a complete, portable semantic representation of information—that could serve as the foundation for future breakthroughs in AI safety, alignment, and economic democratization.
Modern AI systems excel at recognizing and reproducing established patterns. However, when tasked with describing or transmitting novel epistemic insights—especially those that lie outside mainstream consensus—their responses tend to conform to pre-existing paradigms. In other words, while an AI may detect a novel idea, it often fails to preserve the idea’s unique structure and multifaceted nuances when communicating it further. This loss of novelty is problematic for several interrelated reasons:
2.1 Novelty Detection and Preservation Test
Objective:
Assess whether AI models (e.g., Claude Haiku and Google Gemini) can both detect novel epistemic insights and preserve their unique structure in subsequent reasoning.
Methodology & Details:
Implication:
The experiment reveals a twofold challenge:
2.2 First-Order Mechanism for Modeling Novelty
Objective:
Determine what modifications are necessary to improve the preservation of novelty by quantifying how much of a novel idea is retained versus “flattened” into known patterns.
Methodology & Details:
Implication:
Enhancing the capacity for preserving novelty is critical for both advancing AGI and improving AI alignment. Without the ability to maintain epistemic uniqueness, AI systems risk becoming locked into static, consensus-driven approaches that cannot fully exploit the potential of truly innovative ideas.
Central to my work is the development of a “conceptual space”—a complete semantic representation of information that serves as a portable repository of knowledge. This representation is intended to:
Scaling Problem-Solving via a Unified Conceptual Space
A complete semantic representation makes it possible to visualize how the solution space for any given problem—such as AI alignment—expands non-linearly with additional participants. In traditional settings, individual contributions tend to overlap or remain siloed, resulting in a sub-linear increase in overall problem-solving capacity. By contrast, a decentralized collective intelligence (DCI) platform that leverages a unified conceptual space enables participants to connect disparate insights in synergistic ways. This synergy can drive exponential growth in problem-solving ability, which is essential for tackling complex challenges. If efforts remain confined to consensus-based reasoning that neglects these novel regions, progress on critical issues may stall.
The experiments highlight a concerning trend: if AI systems predominantly rely on consensus-based reasoning, they risk filtering out the very insights that could drive breakthrough improvements. Two major consequences follow:
A notable example of how novelty assessment can be fundamentally flawed is found in the study by Cao, Y., Dong, Y., Kim, M., MacLaren, N. G., Pandey, S., Dionne, S. D., Yammarino, F. J., and Sayama, H. (2025) in npj Complexity. Their paper, “Effects of Network Connectivity and Functional Diversity Distribution on Human Collective Ideation,” attempted to gauge novelty by measuring the degree to which collective ideas aligned with prevailing consensus. In doing so, the study conflated the objective “truth” of an insight—the groundbreaking quality of an idea—with “consensus,” or what is widely accepted.
Why This Matters
Given the significant limitations of traditional, centralized funding models—which concentrate resources in the hands of a few decision-makers and favor short-term, consensus-driven projects—it is imperative to explore alternative funding mechanisms. This is especially relevant when considering that preserving epistemic novelty is crucial for breakthrough innovation. If funding mechanisms are designed solely around consensus-based criteria, they risk perpetuating the same limitations that prevent the detection and preservation of true novelty.
The Role of Alternative Funding Models
Meg Lister of Gitcoin, an expert in the web3 space, has proposed several blockchain-powered funding approaches aimed at overcoming these issues:
By leveraging these alternative funding models, we can create a more decentralized, transparent, and community-driven funding environment. Such an approach is essential not only for supporting DCI platforms but also for ensuring that the economic benefits derived from AI innovations are widely distributed, thereby safeguarding both technological progress and collective well-being.
The experiments detailed here reveal that while current AI models are adept at detecting novel insights, they struggle to preserve the epistemic uniqueness of these ideas. This limitation not only hampers progress toward AGI but also risks locking AI into a consensus-driven mode of operation that may be misaligned with human values. Moreover, if our funding and innovation ecosystems continue to favor narrow, consensus-based approaches, there is a danger that only a select few will capture the economic benefits of AI, further deepening inequality.
It is therefore imperative to develop decentralized collective intelligence (DCI) platforms and alternative funding mechanisms that value and integrate rare, non-consensus insights. Notably, the conceptual space framework—which provides a complete semantic representation of information—offers a pathway toward visualizing and preserving the non-linear expansion of the solution space as more participants contribute. This is critical for solving complex challenges like AI alignment, where traditional approaches that neglect novel regions of the conceptual space may ultimately prove unsolvable.
I welcome discussion and feedback on these experiments and their implications for AI alignment, decentralized collective intelligence, and alternative funding models. Your contributions are essential as we work together to develop methods for preserving epistemic novelty in the age of AGI—methods that empower decentralized problem-solving and secure a future where AI benefits all of humanity, not just a select few.